banner



You Have Practiced Enough To Gain A Reward From This Dummy

J Mach Learn Res. Author manuscript; available in PMC 2013 Jun 3.

Published in final edited form as:

J Mach Learn Res. 2012 Nov; 13(Nov): 3253–3295.

PMCID: PMC3670261

NIHMSID: NIHMS445983

Linear Fitted-Q Iteration with Multiple Reward Functions

Daniel J. Lizotte

David R. Cheriton School of Computer Science, University of Waterloo, Waterloo, ON N2L 3G1, Canada, AC.OOLRETAWU@ETTOZILD

Michael Bowling

Department of Computing Science, University of Alberta, Edmonton, AB T6G 2E8, Canada, AC.ATREBLAU.SC@GNILWOB

Susan A. Murphy

Department of Statistics, University of Michigan, Ann Arbor, MI 48109-1107, USA, UDE.HCIMU@YHPRUMAS

Abstract

We present a general and detailed development of an algorithm for finite-horizon fitted-Q iteration with an arbitrary number of reward signals and linear value function approximation using an arbitrary number of state features. This includes a detailed treatment of the 3-reward function case using triangulation primitives from computational geometry and a method for identifying globally dominated actions. We also present an example of how our methods can be used to construct a real-world decision aid by considering symptom reduction, weight gain, and quality of life in sequential treatments for schizophrenia. Finally, we discuss future directions in which to take this work that will further enable our methods to make a positive impact on the field of evidence-based clinical decision support.

Keywords: reinforcement learning, dynamic programming, decision making, linear regression, preference elicitation

1. Introduction

Within the field of personalized medicine, there is increasing interest in investigating the role of sequential decision making for managing chronic disease (Weisz et al., 2004; McKay, 2009; Kuk et al., 2010). Reinforcement learning methods (Szepesvári, 2010; Sutton and Barto, 1998) are already being used (Pineau et al., 2007; Murphy et al., 2007; Zhao et al., 2009) to analyze Sequential Multiple Assignment Randomized Trials (SMART) (Murphy, 2005). A patient's progression through a SMART is divided into stages, each of which consists of a (typically uniform) random assignment to a treatment, followed by monitoring and recording data on the patient's condition. The patient data collected during each stage are very rich and commonly include several continuous variables related to symptoms, side-effects, treatment adherence, quality of life, and so on. For the ith patient in the trial, we obtain a trajectory of observations and actions of the form

o 1 i , a 1 i , o 2 i , a 2 i , , o T i , a T i , o T + 1 i .

Here, a t i represents the action (treatment) at time t, and o t i represents measurements made of patient i after action a t 1 i and before action a t i . The first observations o 1 i are baseline measurements made before any actions are taken.

To analyze these data using reinforcement learning methods, we must define two functions st (o 1, a 1,…,ot ) and rt (st , at , o t+1) which map the patient's current history to a state representation and a scalar reward signal, respectively. Applying these functions to the data from the ith patient gives a trajectory

s 1 i , a 1 i , r 1 i , s 2 i , a 2 i , r 2 i , , s T i , a T i , r T i .

These redefined data are treated as sample trajectories from a known policy which is typically uniformly random over possible actions. Once we have these, we will view ongoing patient care as a Markov decision process (MDP) (Bertsekas and Tsitsiklis, 1996), and apply batch off-policy reinforcement learning methods to learn an optimal policy that takes a patient state and indicates which action appears to be best in view of the data available. In an MDP, both the state transition dynamics and the reward distributions are assumed to have the Markov property. That is, given the value st of the current state, the distribution of next state S t+1 and current reward Rt is conditionally independent of s j, aj, r j for all j < t. Clearly this can be achieved by including past history in st , but this may not be feasible. For this work, we will assume that practitioners can suggest state features (which may be summaries of history) that are "as good as a complete history" in terms of making predictions about future states and rewards: we want features that are rich enough to provide good predictions about action values, but that are simple enough to allow us to learn from a limited amount of data. In medical domains, we may additionally want the learned policy to be easily interpreted and implemented. The interplay between predictiveness, learnability, and interpretability makes the definition of st a challenging problem that requires a great deal of further investigation, particularly into the consequences of a non-Markov definition of state. However, the question of how st should be defined can be answered at least in part by the data themselves together with expert knowledge and feature/model selection techniques analogous to those used in supervised learning settings (Keller et al., 2006) if we have an adequate definition of rt .

A major difficulty with using trial data in this way is that there is often no obviously correct way to define rt . Indeed, any definition of rt is an attempt to answer the question "What is the right quantity to optimize?"—a question that is driven by the objectives of individual decision makers and cannot be answered by the data alone. There are many reasonable reward functions one could define, since each patient record includes a multi-dimensional measurement of that patient's overall well-being. For example, data often include a measure of the severity of the symptoms the patient is experiencing, as well as a measure of the severity of the side-effects caused by the current treatment. These different dimensions are typically better addressed by some treatments than by others, and therefore the choice of which dimension to use as the reward will affect the resulting learned policy. For example, a policy that minimizes expected symptom level will tend to choose more aggressive drugs that are very effective but that have a more severe side-effect profile. On the other hand, a policy that minimizes expected side-effect measurements will choose drugs that are less effective but that have milder side-effects.

In clinical practice, doctors, patients, and families decide on a treatment by weighing different measures of well-being, like symptoms and side-effects, according to subjective preferences that are not known to us at the time of data analysis. Continuing our example, these preferences may lean more toward symptom reduction or side-effect reduction, depending on the individual decision makers involved, and an appropriate reward function definition should reflect these preferences. In principle, one could elicit these preferences, use them to define the reward function of each individual decision maker, and then learn a policy for that reward function; however accurate preference elicitation can be difficult to achieve, and even when it is possible it can be a time-consuming process for the decision maker. Moreover, this approach is problematic because it does not give a complete picture of the quality of the available actions under different reward choices. Indeed, the decision maker will not know when very small changes in preferences might lead to different actions, or when one action is optimal for a broad range of preferences, or when another action is not optimal for any preference.

Rather than eliciting preference and producing a policy that recommends a single action per state, our "inverse preference elicitation" approach is to first consider all of the actions available at the current state. For each of the actions, we answer the question, "What range of preferences makes this action a good choice?" This provides much richer information about the possible actions at each stage. Furthermore, even if a preference is specified somehow, our methods allow the maker to immediately see if his or her preferences are near a "boundary"—that is, whether a small change in preference can lead to a different recommended action. In this case, according to the data analysis two or more actions perform comparably well, and therefore the final decision could be based on other less crucial considerations such as dosing schedule and difference in cost. We are interested in efficient algorithms that can exactly compute the optimal policy for a range of reward functions to investigate how our choice of reward function influences the optimal policy, and in turn how we can offer more flexible choices among good actions.

2. Related Applications, Existing Methods, and Our Contributions

Our approach can help explore trade-offs in different application domains besides sequential medical decision making as well. In e-commerce, one may wish to trade off short-term profits with increased brand-visibility. In robotics, one may wish to trade-off rapid task completion against wear-and-tear on equipment. Researchers are already considering trading off water reserves versus flood risk in water reservoir control (Castelletti et al., 2010); our approach could provide further insight here. Even within RL itself, our approach could provide a new perspective on trading off achieving high expected reward with avoiding risk, an issue explored by Mannor and Tsitsiklis (2011). Any problem for which it is difficult or undesirable to formulate a single scalar reward to drive decision making could benefit from our approach.

There is wide interest in making use of multiple reward signals for sequential decision making. Gábor et al. (1998) demonstrated that an MDP with multiple reward signals can be well-defined and solved so long as we are given a fixed partial ordering on reward vectors. Mannor and Shimkin (2004) offer a formalism where actions are chosen to ensure that the long-term average reward vector approaches a "target set". The target set induces an ordering (closeness) on reward vectors which drives the agent's actions. Natarajan and Tadepalli (2005) assume that a scalar reward function is constructed by taking a weighted sum of a reward vector, just as we will. They assume that the weights are given, and that the weights will change over time. Their strategy is to learn a dictionary of policies for different weight vectors that should eventually contain policies that work well for many different preferences. They note that "An interesting direction for future research is to investigate the number of different weight vectors needed to learn all the optimal policies within a desired degree of accuracy," which we will address as part of this work. Early work in this direction (Barrett and Narayanan, 2008) explored the problem of simultaneously computing optimal policies for a class of reward functions over a small, finite state space in a framework where the model is known. Subsequent developments were made that focussed on the infinite-horizon discounted setting and black-box function approximation techniques (Castelletti et al., 2010; Vamplew et al., 2011). We extended the approach of Barrett and Narayanan (2008) to the setting with real-valued state features and linear function approximation, which is a more appropriate framework for analyzing trial data (Lizotte et al., 2010). We also introduced an algorithm that is asymptotically more time- and space-efficient than the Barrett & Narayanan approach, and described how it can be directly applied to batch data. We also gave an algorithm for finding the set of all non-dominated actions in the single-variable continuous state setting. This paper builds on our previous work by contributing:

  • A general and detailed development of finite-horizon fitted-Q iteration with an arbitrary number of reward signals and linear approximation using an arbitrary number of state features

  • A detailed treatment of 3-reward function case using triangulation algorithms from computational geometry that has the same asymptotic time complexity as the 2-reward function case

  • A more concise solution for identifying globally dominated actions under linear function approximation, and method for solving this problem in higher dimensions

  • A real-world decision aid example that considers symptom reduction, weight gain, and quality of life when choosing treatments for schizophrenia

3. Background

We begin by defining the mathematical framework for our problem and describing its relationship to the usual MDP formulation. We then discuss how two existing formalisms, Inverse Reinforcement learning and POMDP Planning, relate to our approach.

3.1 Problem Framework

For each patient, we assume we will choose a treatment action at each timepoint t = 1,2,…T, after which they are no longer under our care. In this finite-horizon sequential decision making setting, the optimal policy in general depends on t (Bertsekas and Tsitsiklis, 1996), so we explicitly maintain separate rt, Qt , and Vt functions for each timepoint, and we define QT rT . Furthermore, it is convenient for our purposes to allow the set of possible states 𝒮 t and the set of possible actions 𝒜 t to depend on time. We then designate the learned policy at a particular time point by π r :𝒮 t  → 𝒜 t .

We consider sets of MDPs that all have the same 𝒮 t , 𝒜 t , and state transition dynamics, but whose expected reward functions rt(st,at) have an additional parameter δ. One may think of δ as a special part of state that: i) does not evolve with time, and ii) does not influence transition dynamics. Each fixed δ identifies a single MDP by fixing a reward function, which has a corresponding optimal 1 state-action value function defined in the usual way via the Bellman equation:

Q t ( s t , a t , δ ) = r t ( s t , a t , δ ) + E S t + 1 s t , a t [ max a A t + 1 Q t + 1 ( S t + 1 , a , δ ) ] .

We will also refer to the optimal state value function V t (s t , δ) = max a∈𝒜 t Q t (s t ,a, δ), and the 2 optimal deterministic policy π t (s t , δ) ∈ argmax a∈𝒜 t Q t (s t ,a, δ). The purpose of δ is to represent the preferences of a decision maker: we presume that a decision maker would like to follow an optimal policy πt (st , δ) for the MDP indexed by the value of δ that represents their preferences, that is, the value of δ for which rt(st,at,δ) is a reflection of their reward function.

In order to mathematize the relationship between preference and δ, we define the structure of rt(st,at ,δ) to be

δ = (δ[0], δ[1],, δ[D−1]),

(1)

r t ( s t , a t , δ ) = δ [ 0 ] r t [ 0 ] ( s t , a t ) + δ [ 1 ] r t [ 1 ] ( s t , a t ) + + ( 1 d = 0 D 2 δ [ d ] ) r t [ D 1 ] ( s t , a t ) .

(2)

where d = 0 D 1 δ [ d ] = 1 . Thus rt(st,at,δ) is a convex combination of the basis rewards r t[0], …, r t[D−1], identified by a vector δ of length D that identifies points on the (D − 1)-simplex. The vector δ represents a preference that assigns weight to each of the basis rewards. For example, if δ[d] = 1 for some index d, then rt(st,at,δ) = rt[d](st, at), and the optimal policy πt(st, δ) will choose actions that optimize the expected sum of rewards as determined by rt[d] . Intuitively, the magnitude of δ[d] determines how much πt (st , δ) "cares about" the dth basis reward. Preferences defined by non-linear combinations of reward have been considered in non-sequential settings (e.g., Thall 2008), but such approaches would be much more computationally challenging in the sequential decision making setting in addition to being much more challenging to interpret; we discuss this in Section 7.1. Throughout, we presume our data are trajectories where the ith one takes the form

s 1 i , a 1 i , r 1 [ 0 ] i , r 1 [ 1 ] i , , r 1 [ D 1 ] i , s 2 i , a 2 i , r 2 [ 0 ] i , r 2 [ 1 ] i , , r 2 [ D 1 ] i , , s T i , a T i , r T [ 1 ] i , , r T [ D 1 ] i .

3.2 Related Approaches

Inverse Reinforcement Learning (IRL) (e.g., see Ng and Russell, 2000) comprises a collection of methods that take as input trajectories acquired by observing an expert and then attempt to infer the reward function of that expert. While IRL methods also operate in a setting with unknown rewards, our goal is quite different since we explicitly assume that our data do not come from an expert—in fact actions are often chosen uniformly randomly. Furthermore, we do not attempt to recover the reward function of any particular agent; we will instead attempt to learn the optimal policy for a set of reward functions. IRL methods could be useful if one wanted to attempt to explicitly learn the preference (i.e., the δ) of a decision-maker under our proposed framework; we leave this as potential future work.

Partially Observable Markov Decision Process (POMDP) planning (Kaelbling et al., 1998) comprises a large collection of methods that are designed to learn policies in the face of partial observability, that is, when the current "nominal state" 3 of the system is not observed. In this framework, the agent maintains a belief state (i.e., distribution) over the current nominal state and defines a value function and policy over these belief states. In the simplest setting with k nominal states, the space of possible belief states is the set of vectors on the (k − 1)-simplex, and value-based exact POMDP planners compute the optimal value function for all possible belief states. In effect, the POMDP planner defines and solves the belief MDP in which the belief states are considered to be the observed (continuous) state.

Our goal in this work is to learn the optimal value function and policy for all possible preferences δ, which also happen live on the simplex. The value functions we learn are similar in structure to those of the POMDP planning problem, but there are at least two important differences.

First and foremost, the value functions and policies we learn are functions of preference and of additional state (e.g., patient information) both of which are assumed to be observed. We will see that in our formulation, for any fixed state the optimal value function is piecewise linear in preference. The preference part of the value function has a structure similar to that of the value function of a belief MDP, which is piecewise linear in the belief state.

Second, in POMDP planning, value is always a convex function of belief state, and this property is crucial in the development of exact and approximate (e.g., Pineau et al. 2006, Wang et al. 2006) methods. However, we will show in Section 4.2.4 that because our approach estimates value functions using regression, the Q-functions in our problem are not convex in δ. Since we do not have convexity, we will develop alternative methods for representing value functions in Section 4.

Despite these two differences, it is possible to interpret our definition of preference as a "belief" that the agent/patient is in exactly one of D different hidden "preference states" each corresponding to a single basis reward. We will not approach the problem from this point of view since we prefer the interpretation that each agent (e.g., patient) has a true observable δ and a corresponding reward function given by (2), but there may be applications where the hidden "preference state" interpretation is preferable. In any case, the two differences mentioned above mean that even if we interpret δ as a belief over preference states, standard POMDP methods are not applicable.

4. Fitted-Q Iteration for Multiple Reward Functions

In order to illustrate the intuition behind our approach, we first describe an algorithm for learning policies for all possible δ in the simplest case: a finite state space with D = 2 basis rewards. We then describe how to accomodate linear function approximation with an arbitrary number of features in the D = 2 setting. We then give a generalization to arbitrary D, and we provide an explicit algorithm for the D = 3 case based on methods from computational geometry.

4.1 Optimal Value Functions for All Tradeoffs: Finite State Space, D=2 Basis Rewards

To begin, we assume that the 𝒮 t are all finite, and that state transition probabilities P(s t+1|st , at ) and expected rewards rt[0](st, at),…,rt[0](st, at) are estimated using empirical averages from the data set and "plugged in" where appropriate. From an algorithmic standpoint, in this setting there is no difference whether these quantities are known or estimated in this way. We therefore present our algorithm as though all expectations can be computed exactly.

First, we consider two basis rewards rt[0] and rt[1] and corresponding preferences δ = (δ[0][1]). In this setting, the range of possible reward functions can be indexed by a single scalar δ = δ[1] by defining

r t (s t ,a t , δ) = (1 − δ) ⋅r t[0](s t ,a t ) + δ ⋅r t[1](s t ,a t ).

We will show that the optimal state-action value function Vt (st ,δ) is piecewise-linear in the trade-off parameter δ. Where appropriate, we will use the notation Vt (st , ·) to represent the function of one argument (i.e., of δ) identified by fixing st . (We will use notation Qt (st, at , ·) and rt (st, at ,·) similarly.) We use an exact piecewise-linear representation of the functions Vt (st ,·) for each state and timepoint, which allows us to exactly compute value backups for all δ more efficiently than the point-based representations of Barrett and Narayanan (2008). Our representation also allows identification of the set of dominated actions, that is, the actions that are not optimal for any (st,δ) pair. Value backups for finite state spaces require two operations: maximization over actions, and expectation over future states.

4.1.1 Maximization

We begin at time t = T, the final time point, 4 and describe how to take a collection of functions QT (sT, aT , ·) for all (sT, aT ) and produce an explicit piecewise-linear representation of VT (sT , ·) by maximizing over a T  ∈ 𝒜 T . In Section 4.1.3, we show how this can be accomplished at earlier timepoints t < T using a divide-and-conquer approach.

The Q-function for the last timepoint is equal to the terminal expected reward function rT , which is linear in δ for each state-action pair as defined in (2), so we have QT (sT, aT ,δ) = (1 − δ). r T[0](sT, aT ) + δ· r T[1](sT, aT ) To represent each QT (sT, aT , ·), we maintain a list 5 of linear functions, one for each action, for each sT . Figure 1(a) shows an example Q-function at a fixed state sT at time T for four different actions, three of which are optimal for some δ and one which is not optimal for any δ. The linear function for each action can be represented by a list of tradeoff points (i.e., [0 1]) together with a list of their corresponding values (i.e., [QT (sT, aT , 0) QT (sT, aT , 1)]) at those tradeoff points. Each can also be represented by a point (QT (sT, aT , 0) QT (sT, aT , 1)) in the plane, as shown in Figure 1(b). These two equivalent representations offer an important conceptual and computational insight that is well-established in the multi-criterion optimization literature (Ehrgott, 2005): the set of actions that are optimal for some δ ∈ [0,1] are exactly those actions whose line-representations lie on the upper convex envelope of the Q-functions, and equivalently, whose point-based representations lie on the upper-right convex hull of the set of points in the plane. In general, we can recover the actions that are optimal on any interval [δ, δ′] by finding the upper-right convex hull of the points {(QT (sT, a,δ), QT (sT, aT , δ′)) : a ∈ {1…|A|}}. This equivalence is important because the time complexity of the convex hull operation on n points in two or three dimensions is O(nlogn) (de Berg et al., 2008)—as fast as sorting.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0005.jpg

Computing VT from QT for all δ by convex hull.

We make use of this equivalence to construct our piecewise-linear representation of VT (s, ·). Commonly-used convex hull routines produce output that is ordered, so it is easy to recover the list of actions that are optimal for some δ, along with the values of δ where the optimal action changes. These values are the "knots" in the piecewise-linear representation. We denote the list of knots of a piecewise-linear function f(·) by Δ(f(·)). The output of a convex hull algorithm is an ordered list of points, each of the form (QT (sT, aT , 0) QT (sT, aT , 1)). In this case, the list is [(0.8,0.2) (0.5,0.6) (0.2,0.7)]. We know from the order of this list that the second knot in VT (s,·) (after δ = 0) occurs where the lines represented by (0.8,0.2) and (0.5,0.6) intersect. Thus we can compute that the line represented by (0.8,0.2) is maximal from δ = 0 to δ = 0.43, at which point it intersects the line represented by (0.5,0.6). After finding the knots, we represent the piecewise-linear value function in Figure 1(a) by the ordered knot-list Δ(VT (sT , ·))=[0.00 0.43 0.75 1.00] and value-list [0.80 0.54 0.58 0.70], rather than by the list of points. To recover the policy at this point, we may also retain a list of lists containing the actions that are optimal at each knot: [[1] [1 2] [2 4] [4]]. This allows us to determine the action or actions that are optimal for any segment by taking the intersection of the action lists for the endpoints of the segment. Note that because VT (sT , ·) is a point-wise maximum of convex 6 functions, it is convex.

Our representation allows us to evaluate V T (s T , δ) = max a∈𝒜 T Q T (s T ,a T , δ) efficiently. Because our knot list and value list are ordered, we can use binary search to find the largest knot in VT (sT , ·) that is less than δ. This tells us which linear piece is maximal for δ, so we only need to evaluate this single linear function. Thus computing VT (sT , δ) takes O(log|Δ(VT (sT , ·))|) time, that is, the time for the cost of the search, rather than the O(|Δ(VT (sT , ·))|) time it would take to evaluate all of the linear functions at δ and then take the maximum.

4.1.2 Expectation

We now show how we use our piecewise-linear representation of V T (s T ,  ⋅ ) to efficiently compute a piecewise-linear representation of

Q T−1(s T−1,a T−1,  ⋅ ) =r T−1(s T−1,a T−1,  ⋅ ) +E s T [V T (S T ,  ⋅ )∣s T−1,a T−1]

using the piecewise-linear representation of VT. To do so, we must evaluate conditional expectations of VT over possible future states.

Consider an example with two terminal states sT = 1 and sT = 2. Suppose that the probability of arriving in state j (conditioned on some (s T−1, a T−1)) is given by θ j . Since each VT (j, ·) is linear over the intervals between Δ(VT (j, ·)), these two functions are simultaneously linear over the intervals between Δ(VT (1, ·)) ∪ Δ(VT (2, ·)), and their weighted average is linear over the same intervals. Therefore the expectation

E s T [V T (S T ,  ⋅ )∣s T−1,a T−1] = θ1 ⋅V T (1,  ⋅ ) + θ2 ⋅V T (2,  ⋅ )

is itself a piecewise-linear function of δ with knot-list Δ(VT (1, ·)) ∪ Δ(VT (2, ·)). Since the function r T−1(s T−1, a T−1, ·) is linear, it does not contribute additional knots, so we can compute the piecewise-linear representation of Q T−1(s T−1, a T−1, ·) by computing its value-list at the aforementioned knots. The value list of Q T−1(s T−1, a T−1, ·) is

[ ( r T 1 ( s T 1 , a T 1 , δ ) + j θ j V T ( j , δ ) ) : δ Δ ( V T ( 1 , ) ) Δ ( V T ( 2 , ) ) ] .

Let k j = Δ ( V T ( s j , ) ) . This construction uses O(k 1 + k 2) space and requires O(k 1 + k 2) evaluations of VT . Note that because Q T−1(s T−1, a T−1, ·) is a positive weighted sum of convex functions, it is convex. Figure 2 illustrates this weighted sum operation.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0006.jpg

Computing expectations using unions of piecewise-linear representations. The graphs of VT (1,δ) and VT (2,δ) show the time T value function at two different states, ST = 1 and ST = 2. The graph of Q T−1(s T−1,a T−1,δ) = 0.5 · V T (1,δ) + 0.5VT (2,δ) shows the expected action-value function if the two states are each reached with probability 0.5 when starting from state S T−1 = s T−1 and taking action A T−1 = a T-1, and there is no immediate reward.

We contrast the piecewise-linear representation approach with that of Barrett and Narayanan (2008). The expectation can also be computed using the point-based representation in Figure 1(b): let Ξi be the set of points in the point-based representation of VT (sT , ·). One can compute the pointbased representation of Q T−1(s T−1, a T−1, ·) by constructing a set of points

{ ( r ( T 1 ) [ 0 ] , r ( T 1 ) [ 1 ] ) + θ 1 ( a 1 , b 1 ) + θ 2 ( a 2 , b 2 ) } , where r ( T 1 ) [ δ ] = r T 1 ( s T 1 , a T 1 , δ ) for all ( a 1 , b 1 ) Ξ 1 , ( a 2 , b 2 ) Ξ 2 .

(3)

and then taking the upper-right portion of the convex hull of this set. Barrett and Narayanan (2008) advocate this procedure and prove its correctness; however, they note that the set given in (3) has |Ξ1| |Ξ2| points that must be constructed and fed into the convex hull algorithm. Since ki =|Ξi| + 1, computing the expectation in this way will take O(k1k2) space and O(k1k2 logk1k2) time, which is asymptotically much less efficient than our O(k1 + k2) piecewise-linear representation based approach.

4.1.3 Value Backups for t < T − 1

The maximization procedure described in Section 4.1.1 relies on the linearity of Q T (s T , a T , ·). However, for t < T, we have shown that Q t (s t , a t , ·) is piecewise-linear. We now show how to compute Vt and Qt from Qt+1 by first decomposing Q t+1(s t+1, a t+1, ·) into linear pieces and applying the expectation and maximization operations to each piece. Recall that

Q t (s t ,a t , δ) =r t (s t ,a t , δ) +E S t+1 [V t+1(S t+1, δ)∣s t ,a t ].

We have shown by construction that Q T−1(s T−1, a T−1, ·) is convex and piecewise-linear. In general, Vt (st , ·) is computed by taking a point-wise max over functions Qt (st , at , ·), and Q t−1(s t−1, a t−1, ·) is computed by taking a positive weighted sum of the convex functions r t−1(s t−1, a t−1, ·) and Vt (st , ·). Since both of these operations preserve convexity and piecewise-linearity, it follows by induction that Qt (st , at , ·) is convex piecewise-linear for all t ∈ 1, …, T To compute Qt (st , at , ·), we first identify the knots in E S t+1 [V t+1(S t+1, ·)|st , at ] and store them; this is done in the same way as for t + 1 = T. We then compute the value-list as described above. To compute V t (s t ,  ⋅ ) = max a∈𝒜 t Q t (s t ,a,  ⋅ ), we take the maximum over actions of these piecewise-linear Q-functions using Algorithm 2. First, we decompose the problem of finding max a∈𝒜 t Q t (s t ,a,  ⋅ ) for δ ∈ [0,1] into sub-problems of finding max a∈𝒜 t Q t (s t ,a,  ⋅ ) over intervals of δ where we know the Qt (st , at , ·) are simultaneously linear. The ends of these intervals are given by ⋃ a Δ(Qt (st , at , ·)). We then apply the convex hull algorithm to each of these intervals to recover any additional knots in max a∈𝒜 t Q t (s t ,a t ,  ⋅ ).

The full backup procedure is described in Algorithm 1. In practice, we can avoid running the convex hull algorithm over every interval by checking each interval's end points: if for some action a t we find that Q t ( s t , a t , ) is maximal at both ends of an interval in the current knot-list, then max a Qt (st , at , ·) has no knots inside the interval. Note that though we present our algorithms assuming the reward functions are linear, they will work for piecewise-linear reward functions as well.

4.1.4 Complexity of Qt (st , at , ·) and VT (sT , ·)

Suppose there are ∣𝒮∣ states and ∣𝒜∣ actions at each stage. For any fixed sT , each function QT (sT , i, ·), i = 1..∣𝒜∣ has 2 knots, δ = 0 and δ = 1. Applying Algorithm 2 to produce VT (sT , ·) from these functions generates at most ∣𝒜∣−1 new internal knots, and therefore each VT (sT , ·) has at most (∣𝒜∣−1) + 2 knots. To compute Q T−1(s T−1, a T−1, ·), we take the expectation of VT (sT , ·) over states sT . Since VT (sT , ·) might have different internal knots for every sT , Q T−1(s T−1, a T−1, ·) may have as many as ∣𝒮∣(∣𝒜∣−1) + 2 knots. However, the knots in Q T−1(s T−1, a T−1, ·) will be the same for all sT−1 and aT−1 . Computing VT−1 (sT−1 , ·) using Algorithm 2 adds at most ∣𝒜∣−1 new knots between each pair of existing knots, for a total of (∣𝒜∣−1∣)(∣𝒮∣(∣𝒜∣−1) + 1) + 2. In general, Qt (st , at , ·) may have up to O(∣𝒮∣ Tt ∣𝒜∣ Tt ) may have up to O(∣𝒮∣ Tt ∣𝒜∣(Tt)+1) knots.

To compute Qt (st , at , ·) from rt and Vt+1 , our approach requires O(∣𝒮∣ Tt ∣𝒜∣(Tt)+1) time for each state, for a total of O(∣𝒮∣(Tt)+1∣𝒜∣(Tt)+1) time. In contrast, the approach of Barrett & Narayanan requires O(∣𝒮∣2.(Tt)+1∣𝒜∣2.(Tt)+1log∣𝒮∣2.(Tt)+1∣𝒜∣2.(Tt)+1) time for each of log2∣𝒮∣ pairs of piecewise-linear functions.

Algorithm 1

Value Backup - Finite State Space

Algorithm 2

Max of Piecewise-Linear Functions

4.2 Optimal Value Functions for All Tradeoffs: Linear Function Approximation, D = 2 Basis Rewards

Here, we demonstrate how our previously developed algorithms for value backups over all tradeoffs can be extended to the case where we have arbitrary features of state variables and we use a linear approximation of the Qt functions. Again, we first consider QT and VT , which have the simplest form, and then describe how to compute Qt and Vt at earlier time points. This treatment allows for linear function approximators based on an arbitrary number of state features rather than the single continuous state feature described in Lizotte et al. (2010).

Suppose the expected terminal rewards QT(sT, aT, 0) and QT (sT, aT , 1) are each linear functions of the form Q T ( s T , a T , 0 ) = ϕ s T , a T T β T [ 0 ] and Q T ( s T , a T , 1 ) = ϕ s T , a T T β T [ 1 ] . Here, φ sT, aT is a feature vector 7 that depends on state and action, and the weight vectors β T[0] and β T[1] define the linear relationships between the feature vector and the expectation of the two basis rewards at time T.From Equation (2), we have

Q T ( s T , a T , δ ) = ( 1 δ ) ϕ s T , a T T β T [ 0 ] + δ ϕ s T , a T T β T [ 1 ] , = ϕ s T , a T T [ ( 1 δ ) β T [ 0 ] + δ β T [ 1 ] ] .

(4)

A typical definition of φ sT, aT might include a constant component for the intercept, the measurements contained in sT , the discrete action aT encoded as dummy variables, and the product of the measurements in sT with the encoded aT (Cook and Weisberg, 1999). One could also include other non-linear functions of sT and aT as features if desired. In particular, one could produce exactly the same algorithm described in Section 4.1 by using feature vectors of length ∣𝒮∣ × ∣𝒜∣ that consist of a separate indicator for each state-action pair. In this case the estimated parameters will be precisely the sample averages from Section 4.1. Note from (4) that regardless of the definition of φ aT, sT , the function QT(sT, aT,δ) is linear in δ for any fixed sT, aT .

Recall that we have a set of trajectories of the form

s 1 i , a 1 i , r 1 [ 0 ] i , r 1 [ 1 ] i , s 2 i , a 2 i , r 2 [ 0 ] i , r 2 [ 1 ] i , , s T i , a T i , r T [ 0 ] i , r T [ 1 ] i ,

for i = 1…N with which to estimate the optimal Q functions. In order to estimate QT (sT, aT , 0) and QT (sT, aT , 1), we compute parameter estimates β ^ T [ 0 ] and β ^ T [ 1 ] using ordinary least-squares regression by first constructing a design matrix and regression targets

Φ T = [ ϕ s T 1 , a T 1 T ϕ s T 2 , a T 2 T ϕ s T N , a T N T ] , r T [ 0 ] = [ r T [ 0 ] 1 r T [ 0 ] 2 r T [ 0 ] N ] , r T [ 1 ] = [ r T [ 1 ] 1 r T [ 1 ] 2 r T [ 1 ] N ] .

We then compute parameter estimates

β ^ T [ 0 ] = ( Φ T T Φ T ) 1 Φ T T r T [ 0 ] , β ^ T [ 1 ] = ( Φ T T Φ T ) 1 Φ T T r T [ 1 ] ,

using ordinary least squares. 8 These estimated parameters are then substituted into definition (4), giving Q ^ T ( s T , a T , 0 ) and Q ^ T ( s T , a T , 1 ) . To construct an estimate Q ^ T ( s T , a T , δ ) for arbitrary δ ∈ [0, 1], we could construct a scalar reward using r t[0], r t[1] and δ, and solve for the corresponding β ^ T ( δ ) giving

β ^ T ( δ ) = ( Φ T T Φ T ) 1 Φ T T [ ( 1 δ ) r T [ 0 ] + δ r T [ 1 ] ] ,

but by linearity,

β ^ T ( δ ) = ( 1 δ ) ( Φ T T Φ T ) 1 Φ T T r T [ 0 ] + δ ( Φ T T Φ T ) 1 Φ T T r T [ 1 ] , = ( 1 δ ) β ^ T [ 0 ] + δ β ^ T [ 1 ] .

Thus we only need to solve for β ^ T [ 0 ] and β ^ T [ 1 ] , after which we compute Q ^ T ( s T , a T , δ ) = ϕ s T , a T T β ^ T ( δ ) for any δ by taking convex combinations of these coefficient vectors. Therefore, for t = T, it is straightforward to exactly represent Q ^ T ( s T , a T , δ ) for all δ.

4.2.1 Maximization

For any fixed values of sT and aT , Q ^ T ( s T , a T , ) is a linear function. Therefore, we can use the convex hull method to identify the actions that maximize value at a given sT , and use it to recover the knots in the piecewise-linear V ^ T ( s T , ) .

Figure 3 is an illustration of the pieces of a hypothetical V ^ T ( s T , ) that is a maximization over 10 actions. In this example we define a scalar state feature ψ sT and we construct feature vectors

ϕ s T ,a T = [[1ψ s T ]1 a T =1 [1ψ s T ]1 a T =2 [1ψ s T ]1 a T =10]T,

(5)

where 1 aT =k is the indicator function that takes the value 1 if aT = k and 0 otherwise. Note that this choice of features is equivalent to defining Q ^ using a separate linear regression for each action. Each number in Figure 3 marks the region where that action is optimal at time T. For example, a vertical slice at ψsT = −4 of the value function has three linear pieces where actions 10, 1, and 7 are optimal.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0007.jpg

Diagram of the regions in (ψsT ,δ) space where different actions are optimal at time T. In this example, ψsT ∈ [-6,6].

In the finite state-space case, we explicitly represented VT (sT , ·) separately for each nominal state sT in the MDP in order to allow computation of expectations over terminal states. In contrast, in the linear regression setting, we represent V ^ T ( s T , ) for each observed terminal state s T 1 , , s T N in our data set. That is, we explicitly represent a one-dimensional slice of the value function for each of the s T i by applying Algorithm 2 to construct a piecewise-linear representation for V ^ T ( s T i , ) .

4.2.2 Regression On State Features

At stage T − 1, the regression parameters of our estimate Q ^ T 1 ( s T 1 , a T 1 , δ ) are given by

β ^ T 1 ( δ ) = ( Φ T 1 T Φ T 1 ) 1 Φ T 1 T y ^ T 1 ( δ ) ,

where, for t ∈ {1,…, T – 1}, we define

y ^ t ( δ ) = ( ( 1 δ ) r t [ 0 ] + δ r t [ 1 ] ) + v ^ t + 1 ( δ ) ,

which are the one-step value estimates for time t, where

v ^ t + 1 ( δ ) = [ V ^ t + 1 ( s t + 1 1 , δ ) V ^ t + 1 ( s t + 1 2 , δ ) V ^ t + 1 ( s t + 1 N , δ ) ]

The components of the vector y ^ T 1 ( δ ) are not linear in δ, so for t < T, solving the regression only for δ = 0 and δ = 1 does not completely determine Q ^ t ( s t , a t , δ ) . However, the components of y ^ T 1 ( δ ) are each piecewise-linear in δ. We determine the intervals over which the components are simultaneously linear and then explicitly represent the state-value function at the knots [δ1 δ2 … δ K ] between these intervals. The output accompanying this list of knots is a list of estimated parameter vectors [β T−11) β T−12) … β T−1 K )], each given by β ^ T 1 ( δ k ) = ( Φ T 1 T Φ T 1 ) 1 Φ T 1 T y ^ T 1 ( δ k ) This collection of parameters is analogous to the value list in the finite state-space case, and completely defines Q ^ T 1 ( s T 1 , a T 1 , ) for all sT−1 and aT−1 . As before, we can also retain a list of the optimal actions at each of the knots in order to later recover the policy.

Algorithm 3

Value Backup - Linear Function Approximation, D=2 Basis Rewards

4.2.3 Value Backups for t < T − 1

The procedure for computing the V ^ T ( s T i , ) relies on the linearity of Q ^ T ( s T i , a T , ) , but for t < T, Q ^ t ( s t , a t , ) is piecewise-linear in general. Thus to compute V ^ t ( s t i , ) = max a Q ^ t ( s t i , a , ) for each s t i in our data set, we apply Algorithm 2 for each s t i , using regression to compute Q T−1(s T−1, a T−1, ·) from these functions then proceeds as we did for the t = T − 1 case. The entire procedure is described in Algorithm 3.

4.2.4 Non-Convexity of Q ^ t ( s t , a t , )

For t < T, the resulting Q ^ t ( s t , a t , ) are not necessarily convex in the regression setting, as we alluded to in Section 3.2. To see this, recall that each element of β ^ T 1 ( δ ) is a weighted sum of the piecewise-linear y ^ T 1 ( ) :

β ^ T 1 ( δ k ) = ( Φ T 1 T Φ T 1 ) 1 Φ T 1 T y ^ T 1 ( δ k ) , = w T 1 T y ^ T 1 ( δ )

Here, w T−1 is an 1 × N vector that depends on s and on the data, but does not depend on δ. Elements of w T−1 can be positive or negative, depending on the feature representation used and the particular data set on hand. Therefore, although each element of y ^ T 1 ( ) is a convex, piecewise-linear function, the β ^ t ( ) , and therefore the Q ^ t ( s t , a t , ) may not be convex for t < T. One consequence of this non-convexity is that both the algorithm by Barrett and Narayanan (2008), as well as important algorithms from the POMDP literature (e.g., Pineau et al., 2003) that operate on convex piecewise-linear value functions, cannot represent the function Q ^ t ( s t , a t , ) for t < T.

4.2.5 Complexity of Q ^ t ( s t , a t , ) and V ^ t ( s t , )

Suppose there are N trajectories and ∣𝒜∣ actions at each time point. For any fixed sT and aT , the final learned Q-function Q ^ T ( s T , a T , ) has two knots, one at δ = 0 and one at δ = 1. The terminal value function V ^ T ( s T i , ) is constructed at each of N points in state space by applying Algorithm 2 to the Q ^ T ( s T i , a , ) for each observed terminal state s T 1 , s T 2 , , s T N in 𝒟. Each resulting V ^ T ( s T i , ) has at most ∣𝒜∣−1 new internal knots, and therefore each has at most (∣𝒜∣−1) + 2 knots in total. To compute Q ^ T 1 ( s T 1 , a , ) , we use regression with targets constructed from the N value function estimates V ^ T ( s T i , ) . In general, the knots for each V ^ T ( s T i , ) will be unique. Thus each Q ^ T 1 ( s T 1 , a , ) , whose knots are the union of the knots of the V ^ T ( s T i , ) , will have at most N ⋅ (∣𝒜∣−1) + 2 knots. Computing V ^ T 1 ( s T 1 i , ) using Algorithm 2 adds at most ∣𝒜∣−1 new knots between each pair of knots in the union, for a total of (∣𝒜∣−1∣)(N ⋅ (∣𝒜∣−1) + 1) + 2 knots. In general, Q ^ t ( s , a , ) may have up to O(N Tt ∣𝒜∣ Tt ) knots, and V ^ t ( s , ) may have up to O(N Tt ∣𝒜∣(Tt)+1) knots. To compute the expectation described in Section 4.2.2 at time t, our approach requires O(N Tt ∣𝒜∣(Tt)+1) for each trajectory, for a total of O(N (Tt)+1∣𝒜∣(Tt)+1) time.

4.3 Optimal Value Functions for All Tradeoffs: Linear Function Approximation, D > 2 Basis Rewards

We have seen that, for D = 2 basis rewards, Q ^ t ( s t , a t , ) = ϕ s t , a t T β ^ ( ) is continuous and piecewise-linear, but not convex. This remains true for D reward functions and D tradeoffs δ = δ[0],…, δ[D−1], but as D increases, representing Q ^ t ( s t , a t , ) becomes more difficult. In the general case, Q ^ t ( s t , a t , ) and β ^ t ( ) are linear over pieces that are convex polytopes within the space of possible preferences. We prove this below and show how this insight can be used to develop representations of Q ^ t and V ^ t . As in the D = 2 case, we can construct Q ^ t ( s t , a t , ) and V ^ t ( s t , ) for all tT by taking pointwise maximums and pointwise weighted sums of piecewise-linear functions. All proofs are deferred to Appendix A.

Definition 1 (Linear functions over convex polytopes) A function f:ℝ D  → ℝ is linear over a convex polytope ℛ ⊆ ℝ D if

∃ w ∈ ℝ D :f(δ) = δTw∀ δ ∈ ℛ

Since ℛ is a convex polytope, it can be decomposed into a finite collection of simplices 𝒯 i , each with 𝒟 vertices, such that ℛ = U i 𝒯 i (Grünbaum, 1967). Each simplex is itself a convex polytope. For a simplex 𝒯 with vertices δ12,…,δ D , the weight vector w of a linear function f(δ) = δTw defined over 𝒯 can be computed from the values y 1,y 2,…,y D−1 that f takes on the vertices, together with the locations of the vertices themselves. This is accomplished by solving the system of linear equations for w:

[ δ [ 0 ] 1 δ [ 1 ] 1 δ [ D 1 ] 1 δ [ 0 ] 2 δ [ 1 ] 2 δ [ D 1 ] 2 δ [ 0 ] D δ [ 1 ] D δ [ D 1 ] D ] [ w [ 0 ] w [ 1 ] w [ D 1 ] ] = [ y 1 y 2 y D ]

(6)

Thus, a linear function over a convex polytope can be represented as a piecewise-linear function over simplices. 9

Definition 2 (piecewise-linear functions over collections of convex polytopes)

A function f:ℝ D  → ℝ is piecewise-linear over a collection C of convex polytopes if

∀ ℛ ∈ 𝒞∃ w ∈ ℝ D :f(δ) = δTw∀ δ ∈ ℛ.

Algorithm 4

Algorithm sketch for max

Identify the convex polytopes ℛ i where fi is maximal. Each ℛ i is the intersection of ∣𝒜 t ∣ half-spaces.
Decompose each ℛ i into simplices
Evaluate f max at each vertex in each resulting simplex
Recover the ws as needed using Equation (6)

Algorithm 5

Algorithm sketch for sum

Identify the convex polytopes of the form 𝒰 ∩ 𝒱 over which the sum is linear
Decompose each of these polytopes into simplices
Evaluate f sum at each vertex in each resulting simplex
Recover the ws as needed using Equation (6).

Thus we can completely represent a piecewise-linear function as a collection of (ℛ, w) pairs.

Lemma 3 (Max of linear functions) Given a set of functions {f 1, f 2,…, f N} that are all linear over the same convex polytope ℛ, the function

f max(δ) = max(f 1(δ),f 2(δ),,f N (δ))

is piecewise-linear over the collection of convex polytopes 𝒞 = {ℛ1, ℛ2,, ℛ N } given by

R i = R j = 1 N { δ R D : f i ( δ ) f j ( δ ) } , i = 1 . . N .

Note further that U i = 1 N R i = R , and that the function f max is convex (and therefore continuous) on ℛ, since each fi is convex. These properties immediately suggest a strategy for computing a representation of f max described in Algorithm 4. Figure 4 gives a pictorial representation of this strategy, which allows us to perform the max-over-actions portion of a value iteration backup for D > 2. We now address how to perform the weighted-sum-over-states portion of the backup.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0008.jpg

Pictorial representation of taking max of linear functions for D = 3 basis rewards. The first row of triangles represents three linear functions f 1, f 2, and f 3; darker shading indicates higher function values. The second row shows the convex polytopes ℛ i over which f i is maximal, the decomposition of each of these polytopes into simplices 𝒯 i , and their corresponding weight vectors w i . The continuous piecewise-linear function f max is shown at the bottom.

Lemma 4[Sum of piecewise-linear functions] Given two functions, g1 which is piecewise-linear over a collection 𝒞1 of polytopes, and g2 which is piecewise-linear over a collection 𝒞2 of polytopes, their linear combination

f sum(δ) = α1 g 1(δ) + α2 g 2(δ)

is piecewise-linear over the collection

𝒞1+2 = {𝒰 ∩ 𝒱},s.t. 𝒰 ∈ 𝒞1, 𝒱 ∈ 𝒞2.

This property suggests a strategy for representing f sum described in Algorithm 5. Figure 5 gives a pictorial representation of this strategy, which allows us to perform the weighted-sum-over-states portion of a value iteration backup for D > 2.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0009.jpg

Pictorial representation of taking sum of piecewise-linear functions for D = 3 basis rewards. The top row shows the two functions to be added; darker shading indicates higher function values. In the second row, the left diagram shows the pieces over which the sum is linear. The right diagram of the second row shows the resulting continuous piecewise-linear function.

We can now use these strategies to construct a complete algorithm for the D > 2 case. At time T, we have

Q ^ T ( s T , a T , δ ) = ϕ s T , a T T ( δ [ 0 ] β ^ t [ 0 ] + δ [ 1 ] β ^ t [ 1 ] + + δ [ D 1 ] β ^ T ( D 1 ) ) , = δ T w T ,

where

β ^ T [ d ] = ( Φ T T Φ T ) 1 Φ T T r T [ d ] , w T [ d ] = ϕ s T , a T T β ^ T [ d ] .

Thus for any sT,aT , the function Q ^ t ( s t , a t , ) = ϕ s T , a T T β ^ T ( ) is linear over the piece ℛ{δ:δ[d] > 0} ∩ {δ:Σ d δ[d] = 1}, which is the unit (D−1)-simplex. This is because each element of β ^ T [ d ] ( ) is linear in δ. It follows that V ^ T ( s T , ) = max a Q ^ T ( s T , a , ) is piecewise-linear over the sets described in Lemma 3. To represent the stage T value functions V ^ T ( s T , ) , we apply Algorithm 4 to the Q-functions Q ^ t ( s t , a t , ) of each action a for each sT in our data set. Given this value function at time T, we can compute Q ^ T 1 ( , , β ^ T 1 ( δ ) ) by computing each element of β ^ T 1 ( δ ) as the weighted sum of V ^ T ( s T , ) evaluated at the points sT in our data set by repeated application of Algorithm 5. As in the D = 2 case, these weights are given by the columns of the matrix ( Φ T T Φ T ) 1 Φ T T . At this point, note that for any s T−1,a T−1, the function Q ^ T 1 ( s T 1 , a T 1 , ) is piecewise-linear over the same pieces—they are the pieces identified in Lemma 4. Thus to compute V ^ T 1 we can simply apply Algorithm 4 to each of these pieces. Backups to earlier timepoints proceed analogously.

4.3.1 Complexity

Note that the primitive operations required for fitted-Q iteration—pointwise max and pointwise weighted sum—are precisely the same as in the simpler settings discussed earlier, but the functions we are operating on are (D−1)-dimensional.

Suppose there are N trajectories and ∣𝒜∣ actions at each time point. For any fixed sT and aT , the final learned Q-function Q ^ T ( s T , a T , ) has 1 piece ℛ1 corresponding to the unit (D−1)-simplex. The terminal value function V ^ T ( s T i , ) is constructed at each of N points in state space by applying Algorithm 4 to the Q ^ T ( s T i , a T , ) for each observed terminal state s T 1 , s T 2 , , s T N in 𝒟 and each action aT . Each resulting V ^ T ( s T i , ) has at most ∣𝒜∣ pieces ℛ1,, ℛ∣𝒜∣, supposing each action has a piece where it is optimal. To compute Q ^ T 1 ( s T 1 , a , ) , we use regression with targets constructed from the N value function estimates V ^ T ( s T i , ) . In general, the pieces for each V ^ T ( s T i , ) may be unique. Thus each Q ^ T 1 ( s T 1 , a , ) has pieces formed from all possible intersections between pieces of the N different V ^ T ( s T i , ) , so there may be up to ∣𝒜∣ N such pieces. Applying Algorithm 4 again within each of these pieces means that each V ^ T 1 ( s T 1 i , ) may have ∣𝒜∣ N+1 pieces. In general, Q ^ t ( s , a , ) may have up to O ( A Σ i = 1 T t N i ) pieces, and V ^ t ( s , ) may have up to O ( A Σ i = 1 T t N i ) pieces.

A more detailed complexity analysis would depend on how the pieces are represented, and on how Algorithms 4 and 5 are implemented using computational geometry primitives—we have already seen that for D = 2 basis rewards we can do much better than this worst-case bound. Intuitively this is because most of the intersections between pieces of the N different V ^ T ( s T i , ) are in fact empty. A general treatment of implementing Algorithms 4 and 5 is beyond the scope of this paper; however, we now present a detailed algorithm designed for the D = 3 case that is also much less computationally intensive than the above double-exponential bound suggests.

4.4 Optimal Value Functions for All Tradeoffs: Linear Function Approximation, D = 3 Basis Rewards

We now consider the D = 3 case specifically. The first "algorithmic trick" we will use is to represent functions of δ using two rather than three dimensions, that is,

r t (s t ,a t , δ[0], δ[1]) = δ[0] r t[0](s t ,a t ) + δ[1] r t[1](s t ,a t ) + (1 − δ[0] − δ[1])r t[2](s t ,a t ).

This follows from the constraint that Σ i δ[i] = 1. Note that the set of points (δ[0][1]) : δ[0] + δ[1] ≤ 1, δ[0] ≥ 0, δ[1] ≥ 0 is a convex polytope in ℝ2. In fact it is a simplex, and therefore we can represent the linear function QT(sT,aT ,·) by storing the corners of the simplex 𝒯 = [(1, 0)(0, 1)(0, 0)] together with the parameter vectors

β ^ T ( 1 , 0 ) = ( Φ T T Φ T ) 1 Φ T T r t [ 0 ] , β ^ T ( 0 , 1 ) = ( Φ T T Φ T ) 1 Φ T T r t [ 1 ] , β ^ T ( 0 , 0 ) = ( Φ T T Φ T ) 1 Φ T T r t [ 2 ] .

We can compute a weight-vector representation of the function using Equation (6).

Consider two linear functions Q ^ T ( s T , 1 , ) and Q ^ T ( s T , 2 , ) over 𝒯. To take their pointwise maximum, we must identify the pieces over which the maximum is linear, as described in Lemma 3. The boundary of these two pieces is a line in ℝ2. If this line intersects 𝒯, it will divide 𝒯 into the two pieces. If it does not, then one function must be greater than the other over all of 𝒯. Identifying the pieces can be accomplished by finding where (if anywhere) the dividing line given by Q ^ T ( s T , 1 , ) = Q ^ T ( s T , 2 , ) intersects 𝒯; this is illustrated in Figure 6. We represent V ^ T ( s T , ) by recording the pieces ℛ on either side of the dividing line. Each piece is identified by a set of vertices, along with the value of the max at each vertex. (Note that certain vertices will belong to both pieces.) If there are more than 2 actions, we can take further maxes over each identified sub-piece, partitioning as necessary. This completes the max-over-actions step at time T.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0010.jpg

Identifying the pieces over which the max of two linear functions is linear.

To compute Q ^ T 1 ( s T 1 , a T 1 , ) , we compute each element of β ^ T 1 ( ) at each vertex δ by taking a weighted sum over next states of V ^ T ( s T , ) , again with weights given by columns of ( Φ T T Φ T ) 1 Φ T T . From Lemma 4 we know that we need to identify all of the pieces formed by intersecting the linear pieces of the functions to be summed. Naïvely, one might compute the intersection of all pairs of pieces, but for D = 3 basis rewards we can instead use a constrained Delaunay triangulation (CDT), which essentially gives us only the non-empty intersections and does so much more efficiently than enumerating all pairs. Figure 7 gives a schematic diagram of this procedure. The input to a standard Delaunay triangulation algorithm is a list of points in space. The output is a list of simplices (in this case triangles) that partition space and whose vertices come from the input points. The particular triangles chosen satisfy certain properties (de Berg et al., 2008), but the main appeal for our purposes is the algorithm's O(n log n) running time (Chew, 1987), where n is the number of points to be triangulated. A constrained version of the algorithm allows us to additionally specify edges between points that must be present in the output. The constrained version of the algorithm will add points as needed to satisfy this requirement; again Figure 7 illustrates this. The simplices (triangles) will form the pieces for the elements of β ^ T 1 ( ) , which will define our estimates Q ^ T 1 .

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0011.jpg

Computing the sum of three piecewise-linear functions. The three example value functions each have two linear pieces. The boundary between two pieces is shown by a dotted line. We take all of the vertices, plus the boundaries, and give them as input to a constrained Delaunay triangulation procedure. The output is shown.

The output of the CDT algorithm is a set of pieces over which we know the sum of the piecewise-linear functions will be linear. There are in fact more pieces than are strictly necessary, because linear pieces that have more than three vertices (e.g., quadrilaterals) are divided up by the algorithm. Nonetheless, the output is convenient because we can determine the weight vector w for any simplex using Equation (6). Once we have determined these pieces and vertices, we evaluate V ^ T ( s T , ) at each terminal state and each vertex. Each element of β ^ T 1 ( ) is a piecewise-linear function whose pieces are given by the CDT algorithm, and whose values are given by the appropriate weighted sum of V ^ T ( s T , ) evaluated at the vertices. This gives Q ^ T 1 . The max operation to obtain V ^ T 1 can again be achieved by taking the max over each piece of Q ^ T 1 , and so on backward to t = 1. A complete description is given in Algorithm 6

The problem of finding intersections between lines and computing triangulations is well-studied in the field of computational geometry (de Berg et al., 2008). Though these problems may appear trivial, it is very important to avoid situations where there is "ill-conditioning." For example, if we were to use floating point arithmetic to define three lines that should intersect at the same point, we may find that the "intersection point" is different depending on which pair of lines is used to compute it. This can lead to many spurious points and edges being generated as we proceed with value iteration. We take advantage of Cgal, the Computational Geometry Algorithms Library (CGAL, 2011), which is designed specifically to avoid these problems.

Algorithm 6

Value Backup - Linear Function Approximation, D=3 Basis Rewards

4.4.1 Complexity for D = 3

Any triangulation of n points in the plane contains O(n) triangles (Brass, 2005), so the operation Δ Q ^ t constrained _ Delaunay _ Triangulation ( Δ Q ^ t ) increases the size of Δ Q ^ t only linearly. It follows that each Q ^ T 1 ( s T 1 , a , ) has O(N ⋅ ∣𝒜∣) pieces rather than the ∣𝒜∣ N given by the worst case analysis in Section 4.3.1. Therefore Q ^ t ( s , a , ) may have up to O(N Tt ∣𝒜∣ Tt ) pieces, and V ^ t ( s t , ) may have up to O(N Tt ∣𝒜∣(Tt)+1) pieces. Note that these rates are the same as for the D = 2 special case discussed in Section 4.2.5. Intuitively this is because the triangulation of n points in d-dimensional space has O(n [d∕2]), triangles (Brass, 2005), that is, the same asymptotic growth rate for D = 2 (one-dimensional preference space) and D = 3 (two-dimensional preference space).

5. Dominated Actions

The ability to compute Q ^ and V ^ for all preferences achieves our goal of informing the decision maker about the quality of available actions under different preferences, and of informing the decision maker about how the recommended policy changes with preference. In addition to achieving these primary goals, our representations of Q ^ and V ^ allow us to compute whether or not an action is dominated (not optimal for a given state no matter what the preference) and whether it is globally dominated (not optimal for any state-preference pair.) The notion of domination arises in POMDP planning as well, where certain actions may not be optimal for any belief state, but the notion of global domination has no direct analog in the POMDP setting since it is a property of additional observed state s that is not part of a typical POMDP.

The concept of domination is central to the field of multi-criterion optimization (Ehrgott, 2005), and is important in the medical decision making setting because it identifies treatment actions that are not appropriate no matter what the decision maker's preference is. One may also consider actions that are dominated for all patient states in a population of interest, that is, the actions that are globally dominated. Knowing this set of actions would be useful for developing a formulary—a list of treatments that should generally be made available to a patient population.

The general problem of analytically identifying the set of globally dominated actions is difficult, as we will illustrate, but we first provide solutions for low-dimensional problems and discuss the identification of globally dominated actions in higher dimensional settings. Our approach for determining if actions are dominated is to look for certificates of non-domination for each action. A point (st ,at ,δ) where Q ^ t ( s t , a t , δ ) = V ^ t ( s t , δ ) is a certificate that action at is not dominated at state st , and that action at is therefore not globally dominated. 10 All proofs are deferred to Appendix A.

5.1 Finite Case, D = 2 Basis Rewards

We showed in Section 4.1 how to exactly represent the Qt(st,at, ·) and Vt (st , ·) functions for all st and at for D = 2 when the state space is finite by representing them as lists of knots (vertices) and knot-values (vertex-values). In addition to storing this information, we may also store the set of actions that are optimal at each knot, that is, we may store A ( δ ) = { a t : Q t ( s t , a t , δ ) = V t ( s t , δ ) } for each δ in the knot list of Vt (st , ·). Note that 𝒜(δ) may contain more than one action. Suppose δk and δ k+1 are adjacent knots in Δ(Vt (st , ·)). For all δ s.t. δk < δ < δ k+1, we have 𝒜(δ) = 𝒜 k ) ∩ 𝒜 k+1). Thus the set of actions that have a non-domination certificate at state st , is given by

and any actions not in the above union are dominated at st . Note that recording this additional information does not increase the time complexity of the method. It also allows us to find every globally dominated action by computing the above union at each finite state, taking the union of those sets, and identifying actions not present in the union.

5.2 Regression Case, D = 2 Basis Rewards, One State Feature

We now show how to identify all of the globally dominated actions in the linear function approximation setting. We first discuss the case with a single state feature ψsT , D = 2 basis rewards, and the last timepoint T. We also construct feature vectors φsT, asT so that the Q ^ t functions are built using separate regressions for each action; for example see (5). We can then define β ^ t a ( δ ) to be the 2 × 1 sub-vector of β ^ t ( δ ) that aligns with the sub-vector of φst, ast that is non-zero at = a. We also define the matrix B T a = [ β ^ T a ( 0 ) β ^ T a ( 1 ) ] for each action, so that

Q ^ T ( s T , a T , δ ) = [ 1 ψ s T ] B T a [ δ 1 δ ] .

(7)

To find the globally dominated actions, we will search for certificates of non-domination in (ψsT ,δ) space and identify the actions that do not have a certificate. Figure 3 shows an example of this setting. In the example, actions 1, 4, 6, 7, 8, 9, and 10 have regions where they are optimal, and hence certificates of non-domination. Actions 2, 3, and 5 have no certificates, that is, they are not optimal for any combination of ψsT and δ.

Each B T a is a constant given the data and the regression algorithm. The form of (7) clearly shows that Q ^ T ( , a , ) is a bilinear function of ψsT and δ. To analytically identify the set of dominated actions, we analyze the boundaries between the regions where one action has higher value than another. These boundaries occur where QT(·, a 1, ·)= QT(·, a 2, ·) for some actions a 1 and a 2, that is, where

[ 1 ψ s T ] B T a 1 [ δ 1 δ ] = [ 1 ψ s T ] B T a 2 [ δ 1 δ ] ,

which describes the hyperbola in δ and ψsT given by

[ 1 ψ s T ] ( B T a 1 B T a 2 ) [ δ 1 δ ] = 0 .

(8)

Along these boundaries, "triple-points" occur at (ψsT ,δ) points where three or more actions have exactly the same value. At these points, either all of the actions involved are optimal, or none of them are. We now show that if there exists a certificate of non-domination for action a, but there exists no certificate for a on the boundary of the domain of VT (sT , δ), then there exists a certificate for a at a triple-point.

Lemma 5 (Lizotte et al., 2010) If action a is optimal at time T for some point (ψsT ,δ) but is not optimal for any (ψsT ,δ) on the boundary of the domain, then a is optimal for some (ψsT ,δ) that is a triple-point.

From Lemma 5 we know that to find all actions that are optimal for some (ψsT ,δ) we need only check the boundaries and the triple points. The boundaries can be checked using Algorithm 2. (Note that because Q ^ T ( , a , ) is bilinear in δ and in ψsT , we can also use Algorithm 2 to identify for any fixed δ the actions that are optimal for some ψsT .) We can then enumerate the ( A 3 ) triple-points and check them to detect any regions that do not intersect the boundary of the domain, like for example the region where action 1 is optimal in Figure 3 where we have identified the triple-points with white dots. This procedure reveals all actions that are optimal for some (ψsT ,δ), and thereby identifies any actions that are not optimal for any (ψsT ,δ).

To compute the triple points, we must solve the following system of bilinear equations for ψsT and δ:

[ 1 ψ s T ] ( B T a 1 B T a 2 ) [ δ 1 δ ] = 0 , [ 1 ψ s T ] ( B T a 1 B T a 3 ) [ δ 1 δ ] = 0 .

There are many ways of interpreting this system of equations; it describes the intersection of two hyperbolas, as pointed out in our earlier work (Lizotte et al., 2010). We describe a more concise solution here. Note that any solution (ψsT ,δ) must have the property that the vector [1 ψsT ] is orthogonal to the two vectors given by ( B T a 1 B T a 2 ) [ δ 1 δ ] and ( B T a 1 B T a 3 ) [ δ 1 δ ] . Since [1 ψsT ] is two-dimensional, this implies that these two vectors are collinear. Therefore the vector [ δ 1 δ ] must satisfy

( B T a 1 B T a 2 ) [ δ 1 δ ] = λ ( B T a 1 B T a 3 ) [ δ 1 δ ] .

(9)

Equation (9) describes the generalized eigenvalue problem (Golub and Van Loan, 1996). Common software packages can solve for λ and δ. 11 We have described the process of identifying globally dominated actions for bilinear functions; we can immediately extend this algorithm for piecewise bilinear functions by applying it between pairs of knots.

5.3 Regression Case, p State Features, Arbitrary D

Lizotte et al. (2010) conjectured that an analogue of Lemma 5 holds in higher dimensions, and that identifying all globally dominated actions for more state variables and/or reward functions would require computing intersections of surfaces in higher dimensions. We refine this conjecture, and we propose a solution method for finding globally dominated actions for

Q ^ T ( ψ s T , a , δ ) = [ 1 ψ s T T ] B T a δ .

where ψsT is p-dimensional, and δ is the D-dimensional vector of preferences defined as usual. We consider finding non-dominated actions over a "domain of interest," of the form 𝒮 × 𝒟, where 𝒮 is a rectangle in ℝ p , and 𝒟 is a convex subset of valid preferences. To prove our method is correct, we require the following conjecture.

Conjecture 6 If the system of polynomial equations of the form

[ 1 ψ s T T ] ( B T a 1 B T a 2 ) δ = 0 [ 1 ψ s T T ] ( B T a 1 B T a 3 ) δ = 0 [ 1 ψ s T T ] ( B T a 1 B T a k ) δ = 0

(10)

has a finite number of solution points, those points taken together are a continuous vector-valued function of the coefficients of the system.

Conjecture 6 is true for a single polynomial of one variable over the complex plain (Uherka and Sergott, 1977), and holds for the D = 2 case. We believe the conjecture holds because it is known that systems of multivariate polynomials can be reduced to solving a collection of independent problems each involving a single polynomial of one variable. This reduction uses the methods of elimination and extension (Cox et al., 1997).

Proposition 7 Assume Conjecture 6. If there exists a point (ψsT ,δ) in the interior of the domain of interest where action a is optimal, but a is not optimal at any point (ψsT ,δ) where p + D + 1 actions are simultaneously optimal, then there exists a point (ψsT ,δ) on the boundary of the domain of interest where action a is optimal.

We refer to a point where k actions are simultaneously optimal as a "k-tuple point." To find the set of globally non-dominated actions, we first solve (10) and check to see if any of the (p + D + 1)-tuple points are optimal. If so, all of the point's associated actions not globally dominated. The system (10) of polynomial equations can be solved by computer algebra systems or using numerical approximation techniques (Cox et al., 1997; Sturmfels, 2002). By recursively applying the proposition to the boundaries of the original domain, we can ensure that we identify every action that is not globally dominated: first, we find (D + p + 1)-tuple points inside the original (D + p)-dimensional domain of interest and check whether any of these are optimal. We then treat each of the (D + p − 1)-dimensional boundaries as our new domains of interest, and look for (D + p)-tuple points in each of these, and so on until we check each of the 2 D+p zero-dimensional points at the corners of our original domain. Again, we have described the process of identifying globally dominated actions for functions linear in δ; we can immediately extend this algorithm to piecewise-linear functions by applying it within linear regions.

6. Application to Medical Decision Making

An important application of this work is the improvement of the use of sequential medical data for constructing clinical decision support systems. In this section, we briefly discuss how such systems are currently constructed, how preferences are currently addressed in the medical decision making community, and how the methods presented in this paper provide a novel and useful way of incorporating preferences in clinical decision support systems. We then present an example using real data that illustrates how our methods can be used to inform clinical decision making.

6.1 Clinical Decision Support, Evidence-Based Medicine, and Preferences

Currently, most clinical decision support systems are constructed using expert opinion (e.g., Working Group for the Canadian Psychiatric Association and the Canadian Alliance for Research on Schizophrenia, 1998; Miller et al., 1999). Although accumulated clinical experience is invaluable in making good treatment decisions, data-derived scientific evidence is playing an increasingly prominent role. Sackett (1996) state that "The practice of evidence based medicine means integrating individual clinical expertise with the best available external clinical evidence from systematic research."

In order to be effective, any evidence-based decision support system must leave room for individual clinical expertise to inform the final decision. The methods we have presented are able to do this by presenting treatment recommendations in a way that incorporates decision maker preferences. There is extensive literature on "preference elicitation," both within and outside the field of medical decision making. In the medical decision making field, however, preference elicitation is usually done at the population level and used to produce generic clinical guidelines, rather than to make recommendations tailored to individual patients (e.g., Bonnichsen, 2011; Davis et al., 2011). In other fields, preference elicitation is done before presenting any information about the available treatments (Boutilier, 2002; Thall, 2008). It is assumed that preference elicitation is able to reliably extract the preferences of the decision maker; in our setting, preference elicitation would attempt to find the δ that represents the preference of a decision maker, run fitted-Q iteration using rt (st ,at ,δ), and recommend a single treatment. This approach leaves no room for individual clinical expertise.

Our methods provide a novel alternative to preference elicitation. Rather than trying to determine which of the uncountable number of possible preferences a user might have, we present, for each available action, the set of preferences for which that action is optimal. That is, we present the policy as a function of preference. We call this approach "inverse preference elicitation" because rather than eliciting a preference and recommending a treatment, we can easily and intuitively show for each treatment the set of preferences that are consistent with its recommendation. By using this approach, the time a user would have spent having his or her preference elicited is now spent directly considering the evidence for how preferences influence recommended treatment.

6.2 Example: CATIE Study

We illustrate inverse preference elicitation using data from the Clinical Antipsychotic Trials of Intervention Effectiveness (CATIE) study. The CATIE study was designed to compare sequences of antipsychotic drug treatments for the care of schizophrenia patients. The full study design is quite complex (Stroup et al., 2003; Swartz et al., 2003); we have therefore chosen a simplified subset of the CATIE data in order to more clearly illustrate the potential of the methods presented in this paper. CATIE was an 18-month study that was divided into two main phases of treatment. Upon entry into the study, most patients began "Phase 1," in which they were randomized to one of five possible treatments with equal probability: olanzapine, risperidone, quetiapine, ziprasidone, or perphenazine. As they progressed through the study, patients were given the opportunity at each monthly visit to discontinue their Phase 1 treatment and begin "Phase 2" on a new treatment. The set of possible Phase 2 treatments depended on the reason for discontinuing Phase 1 treatment. If the Phase 1 treatment was deemed to be ineffective at reducing symptoms, then their Phase 2 treatment was chosen randomly as follows: clozapine with probability 1/2, or uniformly randomly from the set {olanzapine, risperidone, quetiapine} with probability 1/2. If the Phase 1 treatment was deemed to produce unacceptable side-effects, their Phase 2 treatment was chosen uniformly randomly from the set {olanzapine, risperidone, quetiapine, ziprasidone}.

In previous work, we used batch off-policy reinforcement learning to analyze data from this study using a single reward function (Shortreed et al., 2010). We now give two new analyses using the new methods we have presented to examine multiple rewards simultaneously. The basis rewards we consider are measures of symptoms, side-effects, and quality of life.

Symptoms

PANSS For our symptom measurement, we use the Positive and Negative Syndrome Scale (PANSS) which is a numerical representation of the level of psychotic symptoms experienced by a patient (Kay et al., 1987). A higher value of PANSS reflects the presence of more severe symptoms. PANSS is a well-established measure that we have used in previous work on the CATIE study (Shortreed et al., 2010), and is measured for each CATIE patient both at the beginning of the study and at several times over the course of the study. Since having larger PANSS is worse, for our first basis reward r [0] we use 100 minus the percentile of a patient's PANSS at the end of their time in the study. We use the distribution of PANSS at the beginning of the study as the reference distribution for the percentile.

Body Weight

BMI Weight gain is an important and problematic side-effect of many antipsychotic drugs (Allison et al., 1999). Patients in the CATIE study had their Body Mass Index (BMI) (National Institutes of Health., 1998) measured at study intake and several times over the course of the study. Since in this population having a larger BMI is worse, for our second basis reward r [1] we use 100 minus the percentile of a patient's BMI at the end of their time in the study. We use the distribution of BMI at the beginning of the study as the reference distribution for the percentile.

Quality of Life

HQLS Measures of quality of life are intended to assess to what degree a patient's disease is impacting his or her daily life, in terms of a patient's relationships with others, ability to work, emotional state, and ability to carry out daily activities (Cramer et al., 2000). Patients in CATIE were administered the Heinrichs-Carpenter Quality of Life (HQLS) (Heinrichs et al., 1984) scale at intake and repeatedly as they progressed through the study. Since having a higher HQLS is better, for our third basis reward r [2] we use the percentile of a patient's HQLS at the end of their time in the study. We use the distribution of HQLS at the beginning of the study as the reference distribution for the percentile.

6.3 Symptoms versus Weight Gain

We begin by presenting the output of our algorithm for D = 2, using PANSS as described above for r [0], and BMI for r [1]. In Figures 8, 9 and 10 we will present plots of the piecewise linear value function V ^ t ( s t , ) for t = 1,2 and for various representative values of st . When we plot V ^ t ( s t , δ ) as a function of δ, we simultaneously show the learned optimal action using the style and colour of the plotted line. Thus from our plots one can see both the learned value and the learned policy as a function of δ, which enables us to easily see for each action the range of preferences for which that action looks best.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0012.jpg

Multiple rewards analysis showing learned value function and associated learned policy for Phase 2 Efficacy patients. Three value functions are shown, with the associated action chosen by the learned policy, for s 2 = 25, s 2 = 50, and s 2 = 75.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0013.jpg

Multiple rewards analysis showing learned value function and associated learned policy for Phase 2 Tolerability patients. Three value functions are shown, with the associated action chosen by the learned policy, for s 2 = 25, s 2 = 50, and s 2 = 75.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0014.jpg

Multiple rewards analysis showing learned value function and associated learned policy for Phase 1 patients. Three value functions are shown, with the associated action chosen by the learned policy, for s 2 = 25, s 2 = 50, and s 2 = 75.

6.3.1 Phase 2 Analyses

Following the approach of our previous work (Shortreed et al., 2010), we use PANSS at entry to Phase 2 as a continuous state variable s 2 so that we can allow symptom severity to influence optimal action choice. We convert the PANSS scores at entry to Phase 2 into percentiles just as we did for the PANSS reward signal. Furthermore, we learn value functions for the Phase 2 Efficacy patients and the Phase 2 Tolerability patients separately, since these two groups have different sets of possible actions.

We have relatively little data for Phase 2 Efficacy subgroup of patients. Therefore for this subgroup, we combine the actions of giving {olanzapine, risperidone, or quetiapine} into one "not-clozapine" action: A 2 EFF = { CLOZ , not-CLOZ } . The other three drugs are much more similar to each other than they are to clozapine, which is much more toxic and is currently considered a "last resort" for use when symptoms are not effectively managed by other treatments (McDonagh et al., 2010). The feature vectors we use for Stage 2 Efficacy patients are given by

ϕ s 2 , a 2 EFF = [ 1 , 1 a 2 = CLOZ , s 2 , s 2 1 a 2 = CLOZ , 1 TD , 1 EX , 1 ST 1 , 1 ST 2 , 1 ST 3 , 1 ST 4 ] T .

Here, s 2 is the PANSS percentile at entry to Phase 2. Feature 1 a 2=OLAN is an indicator that the action at the second stage was clozapine, as opposed to one of the other treatments. We also have other features that do not influence the optimal action choice but that are chosen by experts to improve the value estimates. 12 1TD is an indicator variable of whether the patient has had tardive dyskinesia (a motor-control side-effect), 1EX indicates whether the patient has been recently hospitalized, and 1ST1 through 1ST4 indicate the type of facility at which the patient is being treated (e.g., hospital, specialist clinic)

For Phase 2 Tolerability patients, the possible actions are A 2 TOL = { OLAN , QUET , RISP , ZIP } , and the feature vectors we use are given by

ϕ s 2 , a 2 TOL = [ 1 , 1 a 2 = OLAN , 1 a 2 = QUET , 1 a 2 = RISP , s 2 , s 2 1 a 2 = OLAN , s 2 1 a 2 = QUET , s 2 1 a 2 = RISP , 1 TD , 1 EX , 1 ST 1 , 1 ST 2 , 1 ST 3 , 1 ST 4 ] T .

Here we have three indicator features for different treatments at Phase 2, 1 a 2=OLAN, 1 a 2=RISP, 1 a 2=QUET, with ziprasidone represented by turing all of these indicators off. Again we include the product of each of these indicators with the PANSS percentile s 2. The remainder of the features are the same as for the Phase 2 Efficacy patients.

Figure 8 shows a plot of the piecewise linear value function V ^ 2 ( s 2 , ) for patients who are in Phase 2 of the study because of a lack of efficacy of the Phase 1 treatment. We plot V ^ 2 ( s 2 , ) for three fixed values of s 2 corresponding to having low PANSS, moderate PANSS, or high PANSS at entry to Phase 2. (These correspond to setting s 2 = 25, s 2 = 50 and s 2 = 75, respectively.) For all three states shown in the plot, the learned policy indicates that clozapine is the best action for a reward based only on PANSS (i.e., for δ = 0), but not-clozapine (olanzapine or risperidone or quetiapine) is best for a reward based only on BMI (i.e., for δ = 1.) We have indicated the values of δ at which the decision changes from one action to the other by dropping down a dotted line. We see that, except for those with a strong preference for controlling BMI, clozapine appears to be the best choice among patients who found their Phase 1 treatment to be ineffective at controlling symptoms. It is clear from the plot that neither action is globally dominated since neither is dominated at any of our example states.

Figure 9 shows a plot of the piecewise linear value function V ^ 2 ( s 2 , ) for patients who are in phase 2 of the study because they could not tolerate the side-effects of their Phase 1 treatment. Again we plot V ^ 2 ( s 2 , ) for three different Phase 2 entry percentiles of PANSS: s 2 = 25, s 2 = 50 and s 2 = 75. Possible treatments are olanzapine, quetiapine, risperidone and ziprasidone. If we use a reward based only on PANSS (i.e., for δ = 0), the learned policy indicates that olanzapine is the best action for those with high or moderate incoming PANSS, and that risperidone is best for those with lower incoming PANSS. Ziprasidone is best for a reward based only on BMI (i.e., for δ = 1) independent of PANSS level. This result agrees with existing research on weight gain associated with these atypical antipsychotics (Allison et al., 1999). Again, we have indicated the values of δ at which the decision changes from one action to another by dropping down a dotted line. In this analysis, we found that quetiapine was globally dominated.

6.3.2 Phase 1 Analysis

For Phase 1 patients, the possible actions are 𝒜1 = {OLAN, PERP, QUET, RISP, ZIP}, and the feature vectors we use are given by

ϕ s 1 , a 1 TOL = [ 1 , 1 a 1 = OLAN , 1 a 1 = PERP , 1 a 1 = QUET , 1 a 1 = RISP , s 1 , s 1 1 a 1 = OLAN , s 1 1 a 1 = PERP , s 1 1 a 1 = QUET , s 1 1 a 1 = RISP , 1 TD , 1 EX , 1 ST 1 , 1 ST 2 , 1 ST 3 , 1 ST 4 ] T .

We have four indicator features for different treatments at Phase 2, 1 a 1=OLAN, 1 a 1=PERP, 1 a 1=QUET, and 1 a 1=RISP, with ziprasidone represented by turing all of these indicators off. We include the product of each of these indicators with the PANSS percentile s 1 at entry to the study, and the remainder of the features are the same as for the Phase 2 feature vectors. (These are collected before the study begins and are therefore available at Phase 1 as well.)

Figure 10 shows a plot of the piecewise linear value function V ^ 1 ( s 1 , ) for patients entering Phase 1 (the beginning) of the study. Again we plot V ^ 1 ( s 1 , ) for three fixed values of s 1 = 25, s 1 = 50 and s 1 = 75. Possible treatments are perphenazine, olanzapine, quetiapine, risperidone and ziprasidone. For all three states shown in the plot, the learned policy indicates that olanzapine is the best action for a reward based only on PANSS (i.e., for δ = 0). Ziprasidone is best for a reward based only on BMI (i.e., for δ = 1), also independent of PANSS level. Again, the result agrees well with existing research (Allison et al., 1999). In this analysis, we found perphenazine and quetiapine to be globally dominated.

6.4 Symptoms vs. Weight Gain vs. Quality of Life

We now present the output of our algorithm for D = 3, using PANSS for r [0], BMI for r [1], and HQLS for r [2]. We use the methods described in Section 4.4 to compute the value functions which map a state st and a three-element preference vector δ to an estimated value. Rather than display the shape of this value function using a surface or contour plot, we have elected to show only the regions of preference space where each action is optimal (i.e., the learned policy) mapped onto an equilateral triangle. This simplifies the presentation, but still allows us to easily see for each action the set of preferences for which that action looks best. 13 In all examples, we show the policy for PANSS percentile st = 50.

6.4.1 Phase 2 Analyses

We use the same state representation as for the D = 2 example. Because we are using the exact same r [0] and r [1] as we did for the D = 2 example as well, we can exactly recover the learned policy of our previous D = 2 analysis from our D = 3 analysis simply by considering all preferences of the form δ = (δ, 1 − δ, 0), that is, the preferences along the upper-left edge of the triangle.

Figure 11 shows the learned policy for patients with s 2 = 50 whose Phase 1 treatment was not efficacious. As in the D = 2 case, we combine the actions of giving {olanzapine, risperidone, or quetiapine} into one "not-clozapine" action. We see that clozapine appears best if the reward is based only on PANSS or on HQLS, and "not-clozapine" appears best only if the preference assigns a relatively large weight to BMI. If we consider the upper-left edge of the triangle where the preferences assign zero weight to HQLS, we get precisely the same policy shown in Figure 8. We also see that clozapine appears best for all preferences that consider only PANSS and HQLS (bottom edge) and for most preferences that consider only HQLS and BMI (upper-right edge.) We hypothesize that this is because there is a strong association between control of schizophrenia symptoms and quality of life; thus treatments that work well for PANSS should also work somewhat well for HQLS. We note however that clozapine occupies a narrower range on the upper-right edge than it does on the upper-left edge. In this analysis it is clear that neither action is globally dominated because neither is dominated at state s 2 = 50.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0015.jpg

Multiple rewards analysis using PANSS, BMI, and HQLS, showing learned policy for Phase 2 Efficacy patients with s 2 = 50.

Figure 12 shows the learned policy for patients with s 2 = 50 whose Phase 1 treatment was not tolerable due to side-effects. Here, we see that olanzapine appears best if the reward is based only on PANSS or on HQLS, and ziprasidone appears best if the preference assigns a relatively large weight to BMI. For intermediate preferences, risperidone appears best. Again if we consider the upper-left edge of the triangle where the preferences assign zero weight to HQLS, we get precisely the same policy shown in Figure 9. We also see that olanzapine appears best for all preferences that consider only PANSS and HQLS (bottom edge.) Note that horizontal lines in the triangle represent sets of preferences where the weight on BMI is held constant. Over much of preference space, these horizontal lines are completely contained within one treatment's optimal region, meaning that given a weight for BMI, the policy usually does not depend on the relative preference for PANSS versus HQLI. We hypothesize again that this is because there is a strong association between symptom control and quality of life. In this analysis, we found that quetiapine was globally dominated.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0016.jpg

Multiple rewards analysis using PANSS, BMI, and HQLS, showing learned policy for Phase 2 Tolerability patients with s 2 = 50.

6.4.2 Phase 1 Analysis

Figure 13 shows the learned policy for patients with s 1 = 50. Again we see that ziprasidone appears best for preferences that assign a large importance to BMI, and olanzapine appears best for other preferences. Again if we consider the upper-left edge of the triangle where the preferences assign zero weight to HQLS, we get precisely the same policy shown in Figure 10. Interestingly, the region where ziprasidone appears best increases as we decrease the importance of PANSS, indicating it may be preferable for patients who are more concerned with weight control and quality of life than with very tight control of symptoms. In this analysis, we found that quetiapine, risperidone, and perphenazine were dominated at our example state s 1 = 50, but we found no action to be globally dominated.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0017.jpg

Multiple rewards analysis using PANSS, BMI, and HQLS, showing learned policy for Phase 1 patients with s 1 = 50.

6.5 Limitations

We note that unlike our previous work using this data, this analysis does not attempt to remove bias induced by missing data, nor does it provide measures of uncertainty for the learned policy (Shortreed et al., 2010). Both of these limitations indicate important directions for future work, as we discuss below.

7. Discussion and Future Work

The methods we have presented comprise a crucial first step towards a data analysis method that can be deployed in clinical decision support systems. However, there are challenges that remain to be addressed.

7.1 The Meaning of Rewards and the Effect of Scaling

Consider an analysis with D = 2 basis rewards at its final time point t = T. For a preference of δ = 0.5, two actions a 1 and a 2 for which 0.5r [0](s T ,a 1)+0.5r [1](s T ,a 1) = 0.5r [0](s T ,a 2)+0.5r [1](s T ,a 2) are indistinguishable. One can think of the preference as setting an "exchange rate" for r [0] and r [1]: in this case, the basis rewards can be exchanged at a one-to-one rate and our happiness with the overall result of an action remains the same. For δ = 0.75, the two actions would be indistinguishable if 0.25r [0](s T ,a 1)+0.75r [1](s T ,a 1) = 0.25r [0](s T ,a 2)+0.75r [1](s T ,a 2), meaning that the loss of one unit of r [1] would have to be compensated by a gain of three units of r [0] in order for the actions to be considered equivalent. The stronger the preference for r [1], the more units of r [0] we need in order to "make up" for the loss a unit of r [1]. Note that this interpretation would not be possible had we chosen to define reward as a non-linear function of preference.

In our example analysis, we chose to convert all of the rewards to percentiles before using them; thus we interpret a preference of δ = 0.5 to mean that the "exchange rate" is one-to-one for percentiles of PANSS and percentiles of BMI. The exchange rate at δ = 0.5 could be shifted however by first multiplying one or both rewards by a constant factor before using them in our algorithm. If we fed 2 · BMI into our algorithm as r [1], the exchange rate at δ = 0.5 would be two percentiles of PANSS equals one percentile of BMI, and the preference at which the policy changes from one action to the other in Figure 9, for example, would shift to the left. The ordering of recommended treatments (olanzapine for lowest δ, risperidone for moderate δ, ziprasidone for high δ) would remain the same, however.

A more extreme version is illustrated in Figure 14, where we use 50 · BMI as one basis reward. In this analysis, the "exchange rate" for δ = 0.5 is one unit of BMI equals 50 units of PANSS. Note that there are still three non-dominated actions, but the regions where two of them are optimal are now very small and "compressed" into an area very near δ = 0. This illustrates a potential pitfall: if the exchange rate represented by δ = 0.5 is not "moderate," resulting decision aids will be at best unhelpful and at worst misleading. However, it also supports the use of the exact algorithms we have presented: even if the rewards are poorly scaled, the set of non-dominated actions remains the same, and they retain their ordering according to delta.

An external file that holds a picture, illustration, etc.  Object name is nihms-445983-f0018.jpg

Multiple rewards analysis showing learned value function and associated learned policy for Phase 2 Tolerability patients, using 50·BMI as one basis reward. Note how scaling the BMI reward affects the regions where different actions are optimal. (Compare with Figure 9.)

Note that if the region where an action is optimal is very small, a naïve grid-search over δ may not detect it. For example, if we ask try to determine in the Figure 14 example which treatments are optimal near a preference of δ = 0 by checking nearby δ, we may miss risperidone. On the other hand, the exact methods we have presented will correctly recognize that the risperidone is non-dominated. A practitioner using our methods might then wish to change the analysis by rescaling one or more of the basis rewards. The nature of this rescaling will of course depend on the application at hand; we intend to formalize this problem in future work.

7.2 Value Function and Policy Approximations

We have shown that the complexity of constructing the exact value function is potentially exponential in the time horizon of the problem. However, we have also shown in our example that although the value function may be very complex, the learned policy may still be very simple. Figure 10 illustrates this: each faint triangle in the figure represents a linear piece of the value function. Though there are many pieces, by and large adjacent pieces recommend the same action. This reflects a large-scale smoothness in the Q-functions, and suggests that a simple, smooth function might approximate the piecewise-linear Q-functions very well while reducing computational cost. Some existing algorithms for POMDPs that approximate the value function (e.g., Pineau et al. 2006, Wang et al. 2006) may be useful, but novel modifications will be needed to use these approximations in our setting. Another class of approximations introduced by Poupart and Boutilier (2002) focuses on compressing the state space of the POMDP. As we discussed in Section 3.2, the number of states in a POMDP roughly corresponds to the number of basis rewards considered by our method. Thus, these methods may lead to a way of computing a simplified or "compressed" view of preferences when the number of basis rewards is large, which could be used both to reduce computational cost and to help users better understand their preferences.

7.3 Measures of Uncertainty and Similar Q-values

In our example, almost all preferences are associated with exactly one optimal action. In practice, it may make more sense to recommend more than one action for a particular preference if the Q-values of those actions are very similar. In the medical setting, one may prefer to allow the physician or patient to break ties if outcomes under different treatments are deemed to be "close." We note two criteria for "closeness" that deserve further study.

Statistical Significance

One reason for recommending a set of treatments arises when there is insufficient evidence that one action is actually superior to another. Ideally, one would like to know if an observed difference in Q-values for different actions is true for the population or if it is present in the data we have simply by chance. The methods we have presented do not provide uncertainty information about the learned value function or policy, and although the algorithm is based on linear regression which itself has well-established methods for statistical inference, it is known that even standard single-reward fitted-Q iteration requires specially tailored statistical methods in order to obtain valid confidence measures (Laber et al., 2009; Shortreed et al., 2010). These methods, based on the bootstrap data re-sampling procedure, can be very computationally intensive even for one reward function; thus it will be crucial to combine them with new approximations to the problem in order to produce analyses in a reasonable amount of time. Methods for mitigating the bias induced by having partially missing data can be computationally intensive as well (Shortreed et al., 2010), and should be investigated concurrently with methods for producing confidence information.

Practical Significance

Even if we have strong statistical evidence that one action has a higher Q-value than another, we may still wish to recommend a set of actions if that difference is too small to be practically meaningful. Methods for mathematizing the idea of a "clinically meaningful difference" are under investigation (Laber et al., 2012); we see promise for integrating them with our methods.

7.4 D > 3 Basis Rewards

To allow more than 3 basis rewards, we need methods that can represent and manipulate piecewise linear functions in higher dimensions. One avenue would be to use extended algebraic decision diagrams, which have been successfully applied to MDPs (Zamani et al., 2012). It is not obvious whether XADD methods provide us with a computationally feasible solution for D > 3, but their use is worthy of future study.

8. Conclusion

We have presented a general and explicit development of finite-horizon fitted-Q iteration with an arbitrary number of reward signals and linear value function approximation using an arbitrary number of state features. This included a detailed treatment of the 3-reward function case using triangulation primitives from computational geometry and a method for identifying globally dominated actions under linear function approximation. We also presented an example of how our methods can be used to construct real-world decision aid by considering symptom reduction, weight gain, and quality of life in sequential treatments for schizophrenia. Finally, we have discussed future directions in which to take this work that will further enable our methods to make a positive impact on the field of evidence-based clinical decision support.

Acknowledgments

We extend our sincere thanks to our reviewers, whose detailed comments have enabled us to substantially improve our work. We acknowledge support from Natural Sciences and Engineering Research Council of Canada (NSERC) and the National Institutes of Health (NIH) grants R01 MH080015 and P50 DA10075. Data used in the preparation of this article were obtained from the limited access data sets distributed from the NIH-supported "Clinical Antipsychotic Trials of Intervention Effectiveness in Schizophrenia" (CATIE-Sz). The study was supported by NIMH Contract N01MH90001 to the University of North Carolina at Chapel Hill. The ClinicalTrials.gov identifier is {"type":"clinical-trial","attrs":{"text":"NCT00014001","term_id":"NCT00014001"}}NCT00014001. This manuscript reflects the views of the authors and may not reflect the opinions or views of the CATIE-Sz Study Investigators or the NIH.

Appendix A. Proofs

Note that Lemmas 3 and 4 are known (or deemed "obvious") in the computational geometry literature, but are proved here for completeness.

A.1 Proof of Lemma 3

Over each ℛ i , f max = fi which is linear. Each ℛ i is an intersection of the convex polytope ℛ with an intersection of half-spaces of the form {δ : fi (δ) ≥ fj (δ)}, which are also convex polytopes. Thus each ℛ i is a convex polytope.

A.2 Proof of Lemma 4

For any point δ in a set 𝒰 ∩ 𝒱 as above, we have g 1(δ) = δTw𝒰 and g 2(δ) = δTw𝒱. Therefore, for such δ, we have

α 1 g 1 ( δ ) + α 2 g 2 ( δ ) = α 1 ( δ T w U ) + α 2 ( δ T w V ) , = δ T ( α 1 w U + α 2 w V ) , = δ T w U V .

Therefore within each set given by the intersections above, both g 1 and g 2 are linear.

A.3 Proof of Lemma 5

Suppose a is optimal for some (ψsT ,δ) in the domain but is not optimal for any (ψsT ,δ) on the boundary of the domain. Further suppose that a is not optimal at any triple-point. Then the region where a is optimal must be completely enclosed by the region where a single other action a′ is optimal. However, by Equation (8), the boundary between the regions where a is superior to a′ and vice-versa is a hyperbola composed of two sets (sheets) that are each continuous and have infinite extent in both ψsT and δ. The set must therefore intersect the boundary of the domain of (ψsT ,δ) and thus there must exist a certificate for a on the boundary. Thus we have a contradiction.

A.4 Proof of Proposition 7

Assume there is a region inside the domain where a is optimal. Assume a is not optimal at any (p + D + 1)-tuple point. Since a is not optimal at a point where p + D + 1 actions are optimal, the region where a is optimal must have on its boundary points where k actions are simultaneously optimal for some k < p + D + 1. Choose the maximum k for which this is true. This boundary is defined by a system of k − 1 polynomial equations on p + D variables of the form (10); call the variables ζ1, ζ2,…,ζp+D . Since we assume the region where a is optimal is in the interior of the domain, there exists an interior point ζ* that is a solution to the system of equations. Create a new system of k − 1 equations and k − 1 unknowns by fixing the last (p + D) − (k − 1) variables to ζ k = ζ k , ζ k + 1 = ζ k + 1 , ζ p + D = ζ p + D . The point ( ζ 1 , , ζ k 1 ) is a solution to this reduced system. Suppose the solution of the reduced system is a continuous function of ζ k , which holds if Conjecture 6 is true. Then if we move ζ k toward a boundary from its original value, either we will find a point satisfying the original system with ζ k on the boundary and the remaining variables in the interior of the domain, or another variable or variables will reach its boundary first, and the remainder of the variables will be in the interior of the domain. In either case, there exists a point on the boundary of the domain of interest where action a is optimal.

Footnotes

1In this work, most Q- and V-functions are either optimal or estimates of optimal. We omit the usual *superscript in most cases, and mark estimates with a hat ^.

2The optimal policy may not be unique, but this does not concern us.

3The term "nominal state" is used to denote the actual unobserved state of the system, so as to distinguish it from the "belief state," which comprises the agent's current beliefs about the system.

4We will write QT rather than Q t=T in this section, and similarly write s T, 𝒜 T , etc.

5We denote an ordered list with objects a,b,c by [a b c].

6The QT (sT ,aT ,·) are each linear, and therefore convex.

7In statistical terms, ϕ s T , a T T represents a row in the design matrix of the linear model.

8We could also use ridge regression with a fixed ridge parameter. All of our techniques immediately apply in this case as well since the parameter estimates remain piecewise linear in δ.

9Equation (6) has a unique solution only if the determinant of the matrix in the equation is non-zero, that is, only if there are no collinearities in the vertices of 𝒯.

10Note that in this work we determine which actions are estimated to be dominated, since we are using estimates Q ^ t and V ^ t to make this determination. Assessing our confidence that an action is truly non-dominated based on available data will require incorporation of appropriate statistical methods (Laber et al., 2009).

11In practice one solves ( B T a 1 B T a 2 ) x = λ ( B T a 1 B T a 3 ) x and then projects x onto the subspace x [1] = 1 − x [0] by dividing it by x [0] + x [1].

12See Section 4.2 by Shortreed et al. (2010) for a more thorough discussion of these kinds of features. When we display value functions and learned policies in our examples, we set all of these indicators to 0 since they are not needed by the learned policy to select actions in the future.

13In addition to the policy, we have indicated the linear regions produced by the Delaunay triangulations using faint lines in order to give a sense of their complexity.

Contributor Information

Daniel J. Lizotte, David R. Cheriton School of Computer Science, University of Waterloo, Waterloo, ON N2L 3G1, Canada, AC.OOLRETAWU@ETTOZILD.

Michael Bowling, Department of Computing Science, University of Alberta, Edmonton, AB T6G 2E8, Canada, AC.ATREBLAU.SC@GNILWOB..

Susan A. Murphy, Department of Statistics, University of Michigan, Ann Arbor, MI 48109-1107, USA, UDE.HCIMU@YHPRUMAS.

References

  • Allison DB, Mentore JL, Heo M, Chandler LP, Cappelleri JC, Infante MC, Weiden PJ. Antipsychotic-induced weight gain: A comprehensive research synthesis. American Journal of Psychiatry. 1999 Nov;156:1686–1696. [PubMed] [Google Scholar]
  • Barrett L, Narayanan S. Learning all optimal policies with multiple criteria; Proceedings of the 25th International Conference on Machine Learning; 2008.pp. 41–47. [Google Scholar]
  • Bertsekas DP, Tsitsiklis JN. Neuro-Dynamic Programming. Athena Scientific; 1996. p. 12. chapter 2.1. [Google Scholar]
  • Bonnichsen O. Elicitation of ostomy pouch preferences: a discrete-choice experiment. Patient. 2011;4(3):163–175. [PubMed] [Google Scholar]
  • Boutilier C. A POMDP formulation of preference elicitation problems; Proceedings of the 18th National Conference on Artificial Intelligence; 2002.pp. 239–246. [Google Scholar]
  • Brass P. On the size of higher-dimensional triangulations. In: Goodman JE, Pach J, Welzl E, editors. Combinatorial and Computational Geometry. Vol. 52. MSRI Publications; 2005. pp. 147–152. [Google Scholar]
  • Castelletti A, Galelli S, Restelli M, Soncini-Sessa R. Tree-based reinforcement learning for optimal water reservoir operation. Water Resources Research. 2010;46(W06502) [Google Scholar]
  • CGAL . Cgal. Computational Geometry Algorithms Library; 2011. URL http://www.cgal.org. [Google Scholar]
  • Chew LP. Constrained delaunay triangulations; Proceedings of the Third Annual Symposium on Computational geometry, SCG '87; New York, NY, USA. 1987.pp. 215–222. [Google Scholar]
  • Cook RD, Weisberg S. Applied Regression Including Computing and Graphics. Wiley; Aug, 1999. [Google Scholar]
  • Cox DA, O'Shea D, Little JB. Ideals, Varieties and Algorithms: An Introduction to Computational Algebraic Geometry and Commutative Algebra. Springer; 1997. [Google Scholar]
  • Cramer JA, Rosenheck R, Xu W, Thomas J, Henderson W, Charney DS. Quality of life in schizophrenia: A comparison of instruments. Schizophrenia Bulletin. 2000;26(3):659–666. [PubMed] [Google Scholar]
  • Davis CC, Claudius M, Palinkas LA, Wong JB, Leslie LK. Putting families in the center: Family perspectives on decision making and ADHD and implications for ADHD care. Journal of Attention Disorders. 2011 Oct; E-pub ahead of print. [PubMed] [Google Scholar]
  • de Berg M, Cheong O, van Kreveld M, Overmars M. Computational Geometry Algorithms and Applications. 3 edition Springer; 2008. [Google Scholar]
  • Ehrgott M. Multicriteria Optimization. second edition. Springer; 2005. chapter 3. [Google Scholar]
  • Gábor Z, Kalmár Z, Szepesvári C. Multi-criteria reinforcement learning; Proceedings of the 15th International Conference on Machine Learning; 1998.pp. 197–205. [Google Scholar]
  • Golub GH, Van Loan CF. John Hopkins Studies in the Mathematical Sciences. Johns Hopkins University Press; 1996. Matrix Computation. ISBN 0471935271. [Google Scholar]
  • Grünbaum B. Convex Polytopes, volume 221 of Graduate Texts in Mathematics. Springer-Verlag; 1967. ISBN 0387004246. [Google Scholar]
  • Heinrichs DW, Hanlon TE, C WT., Jr. The Quality of Life Scale: An instrument for rating the schizophrenic deficit syndrome. Schizophrenia Bulletin. 1984;10(3):388–398. [PubMed] [Google Scholar]
  • Kaelbling LP, Littman ML, Cassandra AR. Planning and acting in partially observable stochastic domains. Artificial Intelligence. 1998;101(12):99–134. [Google Scholar]
  • Kay SR, Fiszbein A, Opfer LA. The Positive and Negative Syndrome Scale (PANSS) for schizophrenia. Schizophrenia Bulletin. 1987;13(2):261–276. [PubMed] [Google Scholar]
  • Keller PW, Mannor S, Precup D. Automatic basis function construction for approximate dynamic programming and reinforcement learning; Proceedings of the 23rd International Conference on Machine learning, ICML '06; 2006.pp. 449–456. [Google Scholar]
  • Kuk AY, Li J, Rush AJ. Recursive subsetting to identify patients in the STAR*D: a method to enhance the accuracy of early prediction of treatment outcome and to inform personalized care. Journal of Clinical Psychiatry. 2010 Nov;71(11):1502–8. [PubMed] [Google Scholar]
  • Laber EB, Qian M, Lizotte DJ, Murphy SA. Technical Report 50. Univ. of Michigan Statistics Department; 2009. Statistical inference in dynamic treatment regimes. [Google Scholar]
  • Laber EB, Lizotte DJ, Ferguson B. Set-valued dynamic treatment regimes for competing outcomes. 2012. arXiv:1207.3100v2 [stat.ME] [PMC free article] [PubMed] [Google Scholar]
  • Lizotte DJ, Bowling M, Murphy SA. In: Fürnkranz J, Joachims T, editors. Efficient reinforcement learning with multiple reward functions for randomized controlled trial analysis; Proceedings of the 27th International Conference on Machine Learning; 2010; Haifa, Israel: Omnipress; Jun, pp. 695–702. [Google Scholar]
  • Mannor S, Shimkin N. A geometric approach to multi-criterion reinforcement learning. Journal of Machine Learning Research. 2004;5:325–260. [Google Scholar]
  • Mannor S, Tsitsiklis JN. Mean-variance optimization in Markov decision processes; Proceedings of the 28th International Conference on Machine Learning; 2011.pp. 177–184. [Google Scholar]
  • McDonagh MS, Peterson K, Carson S, Fu R, Thakurta S. Technical report. Oregon Health & Science University; Jul, 2010. Drug class review: Atypical antipsychotic drugs. Drug Effectiveness Review Project, Update 3. [Google Scholar]
  • McKay JR. Treating Substance Use Disorders with Adaptive Continuing Aftercare. American Psychological Association; 2009. [Google Scholar]
  • Miller AL, Chiles JA, C JK, et al. The Texas Medication Algorithm Project (TMAP) schizophrenia algorithms. Journal of Clinical Psychiatry. 1999;60:649–657. [PubMed] [Google Scholar]
  • Murphy SA. An experimental design for the development of adaptive treatment strategies. Statistics in Medicine. 2005;24:1455–1481. [PubMed] [Google Scholar]
  • Murphy SA, Oslin SW, Rush AJ, Zhu J. Methodological challenges in constructing effective treatment sequences for chronic psychiatric disorders. Neuropsychopharmacology. 2007;32:257–262. [PubMed] [Google Scholar]
  • Natarajan S, Tadepalli P. Dynamic preferences in multi-criteria reinforcement learning; Proceedings of the 22nd International Conference on Machine Learning; 2005.pp. 601–608. [Google Scholar]
  • National Institutes of Health . Clinical Guidelines on the Identification, and Treatment of Overweight and Obesity in Adults: The Evidence Report. National Heart, Lung, and Blood Institute; Sep, 1998. NIH Publication no. 98-4083. [PubMed] [Google Scholar]
  • Ng AY, Russell SJ. Algorithms for inverse reinforcement learning; Proceedings of the 17th International Conference on Machine Learning; 2000.pp. 663–670. [Google Scholar]
  • Pineau J, Gordon G, Thrun S. Point-based value iteration: An anytime algorithm for POMDPs; International Joint Conference on Artificial Intelligence; 2003.pp. 1025–1032. [Google Scholar]
  • Pineau J, Gordon G, Thrun S. Anytime point-based approximations for large POMDPs. Journal of Artificial Intelligence Research. 2006;27:335–380. [Google Scholar]
  • Pineau J, Bellemare MG, Rush AJ, Ghizaru A, Murphy SA. Constructing evidence-based treatment strategies using methods from computer science. Drug and Alcohol Dependence. 2007 May;88(Suppl 2):S52–S60. [PMC free article] [PubMed] [Google Scholar]
  • Poupart P, Boutilier C. Value-directed compression of POMDPs. Advances in Neural Information Processing Systems. 2002;15:1547–1554. [Google Scholar]
  • Sackett DL. Evidence-based medicine: What it is and what it isn't. British Medical Journal. 1996;312(7023):71–72. [PMC free article] [PubMed] [Google Scholar]
  • Shortreed S, Laber EB, Lizotte DJ, Stroup TS, Pineau J, Murphy SA. Informing sequential clinical decision-making through reinforcement learning: an empirical study. Machine Learning. 2010:1–28. ISSN 0885-6125. [PMC free article] [PubMed] [Google Scholar]
  • Stroup TS, McEvoy JP, Swartz MS, Byerly MJ, Glick ID, Canive JM, McGee M, Simpson GM, Stevens MD, Lieberman JA. The national institute of mental health clinical antipschotic trials of intervention effectiveness (CATIE) project: Schizophrenia trial design and protocol deveplopment. Schizophrenia Bulletin. 2003;29(1):15–31. [PubMed] [Google Scholar]
  • Sturmfels B. Solving Systems of Polynomial Equations; Regional conference series in mathematics; 2002; Published for the Conference Board of the Mathematical Sciences by the American Mathematical Society; ISBN 9780821832516. [Google Scholar]
  • Sutton RS, Barto AG. Reinforcement Learning: An Introduction. MIT Press; 1998. [Google Scholar]
  • Swartz MS, Perkins DO, Stroup TS, McEvoy JP, Nieri JM, Haal DD. Assessing clinical and functional outcomes in the clinical antipsychotic of intervention effectiveness (CATIE) schizophrenia trial. Schizophrenia Bulletin. 2003;29(1):33–43. [PubMed] [Google Scholar]
  • Szepesvári C. Algorithms for Reinforcement Learning. Morgan & Claypool; Jul, 2010. [Google Scholar]
  • Thall PF. Some geometric methods for constructing decision criteria based on two-dimensional parameters. Journal of Statistical Planning and Inference. 2008;138(2):516–527. [PMC free article] [PubMed] [Google Scholar]
  • Uherka DJ, Sergott AM. On the continuous dependence of the roots of a polynomial on its coefficients. The American Mathematical Monthly. 1977;84(5):368–370. ISSN 0002-9890. [Google Scholar]
  • Vamplew P, Dazeley R, Berry A, Issabekov R, Dekker E. Empirical evaluation methods for multiobjective reinforcement learning algorithms. Machine Learning. 2011;84:51–80. ISSN 0885-6125.10.1007/s10994-010-5232-5. [Google Scholar]
  • Wang T, Poupart P, Bowling M, Schuurmans D. Compact, convex upper bound iteration for approximate pomdp planning; Proceedings of the 21st National Conference on Artificial Intelligence; 2006.pp. 1245–1251. [Google Scholar]
  • Weisz JR, Chu BC, Polo AJ. Treatment dissemination and evidence-based practice: Strengthening intervention through clinician-researcher collaboration. Clinical Psychology: Science and Practice. 2004;11(3):300–307. ISSN 1468-2850. [Google Scholar]
  • Working Group for the Canadian Psychiatric Association and the Canadian Alliance for Research on Schizophrenia Canadian clinical practice guidelines for the treatment of schizophrenia. Canadian Journal of Psychiatry. 1998;43(suppl. 2):25–40S. [PubMed] [Google Scholar]
  • Zamani Z, Sanner S, Fang C. Symbolic dynamic programming for continuous state and action MDPs; Proceedings of the 26th AAAI Conference on Artificial Intelligence; 2012; To appear. [Google Scholar]
  • Zhao Y, Kosorok MR, Zeng D. Reinforcement learning design for cancer clinical trials. Statistics in Medicine. 2009;28:3294–3315. [PMC free article] [PubMed] [Google Scholar]

You Have Practiced Enough To Gain A Reward From This Dummy

Source: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3670261/

Posted by: norrisrues1974.blogspot.com

0 Response to "You Have Practiced Enough To Gain A Reward From This Dummy"

Post a Comment

Iklan Atas Artikel

Iklan Tengah Artikel 1

Iklan Tengah Artikel 2

Iklan Bawah Artikel