gradient temporal-difference learning with regularized

23
Gradient Temporal-Difference Learning with Regularized Corrections Sina Ghiassian *1 Andrew Patterson *1 Shivam Garg 1 Dhawal Gupta 1 Adam White 12 Martha White 1 Abstract It is still common to use Q-learning and tempo- ral difference (TD) learning—even though they have divergence issues and sound Gradient TD alternatives exist—because divergence seems rare and they typically perform well. However, re- cent work with large neural network learning sys- tems reveals that instability is more common than previously thought. Practitioners face a difficult dilemma: choose an easy to use and performant TD method, or a more complex algorithm that is more sound but harder to tune and all but unex- plored with non-linear function approximation or control. In this paper, we introduce a new method called TD with Regularized Corrections (TDRC), that attempts to balance ease of use, soundness, and performance. It behaves as well as TD, when TD performs well, but is sound in cases where TD diverges. We empirically investigate TDRC across a range of problems, for both prediction and control, and for both linear and non-linear function approximation, and show, potentially for the first time, that Gradient TD methods could be a better alternative to TD and Q-learning. 1. Introduction Off-policy learning—the ability to learn the policy or value function for one policy while following another—underlies many practical implementations of reinforcement learning. Many systems use experience replay, where the value func- tion is updated using previous experiences under many dif- ferent policies. A similar strategy is employed in asyn- chronous learning systems that use experience from several different policies to update multiple distributed learners (Es- peholt et al., 2018). Off-policy updates can also be used to * Equal contribution 1 Amii, Department of Computing Science, University of Alberta. 2 DeepMind, Alberta. Correspondence to: Sina Ghiassian <[email protected]>, Andrew Patterson <[email protected]>. Proceedings of the 37 th International Conference on Machine Learning, Vienna, Austria, PMLR 108, 2020. Copyright 2020 by the author(s). learn a policy from human demonstrations. In general, many algorithms attempt to estimate the optimal policy from sam- ples generated from a different exploration policy. One of the most widely-used algorithms, Q-learning—a temporal difference (TD) algorithm—is off-policy by design: simply updating toward the maximum value action in the current state, regardless of which action the agent selected. Both TD and Q-learning, however, have well documented convergence issues, as highlighted in the seminal counterex- ample by Baird (1995). The fundamental issue is the combi- nation of function approximation, off-policy updates, and bootstrapping: an algorithmic strategy common to sample- based TD learning and Dynamic Programming algorithms (Precup, Sutton & Dasgupta, 2001). This combination can cause the value estimates to grow without bound (Sutton & Barto, 2018). Baird’s result motivated over a decade of research and several new off-policy algorithms. The most well-known of these approaches, the Gradient TD methods (Sutton et al., 2009), make use of a second set of weights and importance sampling. Although sound under function approximation, these Gradi- ent TD methods are not commonly used in practice, likely due to the additional complexity of tuning two learning rate parameters. Many practitioners continue to use un- sound approaches such as TD and Q-learning for good reasons. The evidence of divergence is based on highly contrived toy counter-examples. Often, many large scale off-policy learning systems are designed to ensure that the target and behaviour policies are similar—and therefore less off-policy—by ensuring prioritization is mixed with random sampling (Schaul et al., 2016), or frequently syncing the actor policies in asynchronous architectures (Mnih et al., 2016). However, if agents could learn from a larger variety of data streams, our systems could be more flexible and potentially more data efficient. Unfortunately, it appears that current architectures are not as robust under these more aggressive off-policy settings (van Hasselt et al., 2018). This results in a dilemma: the easy-to-use and typically effective TD algorithm can sometimes fail, but the sound Gradient TD algorithms can be difficult to use. There are algorithms that come close to achieving conver- gence and lower variance updates without the need to tune multiple stepsize parameters. Retrace (Munos et al., 2016)

Upload: others

Post on 01-Jan-2022

10 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Sina Ghiassian * 1 Andrew Patterson * 1 Shivam Garg 1 Dhawal Gupta 1 Adam White 1 2 Martha White 1

AbstractIt is still common to use Q-learning and tempo-ral difference (TD) learning—even though theyhave divergence issues and sound Gradient TDalternatives exist—because divergence seems rareand they typically perform well. However, re-cent work with large neural network learning sys-tems reveals that instability is more common thanpreviously thought. Practitioners face a difficultdilemma: choose an easy to use and performantTD method, or a more complex algorithm that ismore sound but harder to tune and all but unex-plored with non-linear function approximation orcontrol. In this paper, we introduce a new methodcalled TD with Regularized Corrections (TDRC),that attempts to balance ease of use, soundness,and performance. It behaves as well as TD, whenTD performs well, but is sound in cases whereTD diverges. We empirically investigate TDRCacross a range of problems, for both predictionand control, and for both linear and non-linearfunction approximation, and show, potentially forthe first time, that Gradient TD methods could bea better alternative to TD and Q-learning.

1. IntroductionOff-policy learning—the ability to learn the policy or valuefunction for one policy while following another—underliesmany practical implementations of reinforcement learning.Many systems use experience replay, where the value func-tion is updated using previous experiences under many dif-ferent policies. A similar strategy is employed in asyn-chronous learning systems that use experience from severaldifferent policies to update multiple distributed learners (Es-peholt et al., 2018). Off-policy updates can also be used to

*Equal contribution 1Amii, Department of Computing Science,University of Alberta. 2DeepMind, Alberta. Correspondenceto: Sina Ghiassian <[email protected]>, Andrew Patterson<[email protected]>.

Proceedings of the 37 th International Conference on MachineLearning, Vienna, Austria, PMLR 108, 2020. Copyright 2020 bythe author(s).

learn a policy from human demonstrations. In general, manyalgorithms attempt to estimate the optimal policy from sam-ples generated from a different exploration policy. One ofthe most widely-used algorithms, Q-learning—a temporaldifference (TD) algorithm—is off-policy by design: simplyupdating toward the maximum value action in the currentstate, regardless of which action the agent selected.

Both TD and Q-learning, however, have well documentedconvergence issues, as highlighted in the seminal counterex-ample by Baird (1995). The fundamental issue is the combi-nation of function approximation, off-policy updates, andbootstrapping: an algorithmic strategy common to sample-based TD learning and Dynamic Programming algorithms(Precup, Sutton & Dasgupta, 2001). This combination cancause the value estimates to grow without bound (Sutton& Barto, 2018). Baird’s result motivated over a decade ofresearch and several new off-policy algorithms. The mostwell-known of these approaches, the Gradient TD methods(Sutton et al., 2009), make use of a second set of weightsand importance sampling.

Although sound under function approximation, these Gradi-ent TD methods are not commonly used in practice, likelydue to the additional complexity of tuning two learningrate parameters. Many practitioners continue to use un-sound approaches such as TD and Q-learning for goodreasons. The evidence of divergence is based on highlycontrived toy counter-examples. Often, many large scaleoff-policy learning systems are designed to ensure that thetarget and behaviour policies are similar—and therefore lessoff-policy—by ensuring prioritization is mixed with randomsampling (Schaul et al., 2016), or frequently syncing theactor policies in asynchronous architectures (Mnih et al.,2016). However, if agents could learn from a larger varietyof data streams, our systems could be more flexible andpotentially more data efficient. Unfortunately, it appearsthat current architectures are not as robust under these moreaggressive off-policy settings (van Hasselt et al., 2018). Thisresults in a dilemma: the easy-to-use and typically effectiveTD algorithm can sometimes fail, but the sound GradientTD algorithms can be difficult to use.

There are algorithms that come close to achieving conver-gence and lower variance updates without the need to tunemultiple stepsize parameters. Retrace (Munos et al., 2016)

Page 2: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

and its prediction variant Vtrace (Espeholt et al., 2018)reduce the variance of off-policy updating, by clipping im-portance sampling ratios. These methods, however, are builton off-policy TD and so still have divergence issues (Touatiet al., 2018). The sound variants of these algorithms (Touatiet al., 2018), and the related work on an algorithm calledABQ (Mahmood, Yu & Sutton, 2017), maintain some ofthe variance reduction, but rely on Gradient TD to obtainsoundness and so inherit the issues therein—the need to tunemultiple stepsize parameters. Linear off-policy predictioncan be reformulated as a saddlepoint problem, resulting inone time-scale, true gradient descent variant of the GTD2algorithm (Mahadevan et al., 2014; Liu et al., 2015; Liuet al., 2016). The Emphatic TD algorithm achieves con-vergence with linear function approximation and off-policyupdates using only a single set of weights and thus onestepsize parameter (Sutton et al., 2016). Unfortunately,high variance updates reduce the practicality of the method(White & White, 2016). Finally, Hybrid TD algorithms(Hackman, 2012, White & White, 2016) were introducedto automatically switch between TD updates when the datais on-policy, and gradient-style updates otherwise, thus en-suring convergence. In practice these hybrid methods aremore complicated to implement and can have stability issues(White & White, 2016).

In this paper we introduce a new Gradient TD method, calledTD with Regularized Corrections (TDRC). With more regu-larization, the algorithm acts like TD, and with no regular-ization, it reduces to TD with gradient Corrections (TDC).We find that for an interim level of regularization, TDRC ob-tains the best of both algorithms, and is not sensitive to thisparameter: a regularization parameter of 1.0 was effectiveacross all experiments. We show that our method (1) outper-forms other Gradient TD methods overall across a varietyof problems, and (2) matches TD when TD performs wellwhile maintaining convergence guarantees. We demonstratethat TDC frequently outperforms the saddlepoint variantof Gradient TD, motivating why we build on TDC and theutility of being able to shift between TD and TDC by set-ting the regularization parameter. We then highlight whyTDRC improves so significantly on TDC, by examiningTDC’s sensitivity to its second stepsize. We conclude witha demonstration in control, with non-linear function approx-imation, showing that (1) TDC can perform very well insome settings and very poorly in others, and (2) TDRC isalways comparable to Q-learning, and in some cases, ismuch better.

2. BackgroundIn this paper we tackle the policy evaluation problem inReinforcement Learning. We model the agent’s interactionswith its environment as a Markov Decision Process (MDP).

The agent and environment interact continually. On eachtime step t = 0, 1, 2, . . . , the agent selects an action At ∈A in state St ∈ S. Based on the agent’s action At andthe transition dynamics, P : S × A × S → [0, 1], theenvironment transitions into a new state, St+1, and emits ascalar reward Rt+1. The agent selects actions according toits policy π : S ×A → [0, 1]. The main objective in policyevaluation is to estimate the value of a state s, defined as theexpected discounted sum of future rewards under π:

vπ(s)def= Eπ

[Rt+1 + γRt+2 + γ2Rt+3 + · · · |St = s

]

= Eπ[Gt|St = s] , (1)

where γ ∈ [0, 1], Gt ∈ R is called the return, and Eπ is theexpectation taken with respect to future states, actions, andrewards generated by π and P .

In many problems of interest, the agent cannot directlyobserve the state. Instead, on each step the agent observesa featurized representation of the state xt

def= x(St) ∈ Rn,

where n |S|. In this setting, the agent cannot estimate thevalue of each state individually, but must approximate thevalue with a parametric function. In this paper, we focus onthe case of linear function approximation, where the valueestimate v : S × Rn → R is simply formed as an innerproduct between x(s) and a learned set of weights w ∈ Rn

given by v(s,w)def= w>x(s).

Our objective is to adjust wt on each time step to constructa good approximation of the true value: v ≈ vπ. Perhapsthe most well known and successful algorithm for doing sois temporal difference (TD) learning :

δtdef= Rt+1 + γw>t xt+1 −w>t xt

wt+1 ← wt + αtδtxt (2)

for stepsize αt > 0. TD is guaranteed to be convergent un-der linear function approximation and on-policy sampling.

The classical TD algorithm was designed for on-policy learn-ing; however, it can be easily extended to the off-policy set-ting. In on-policy learning, the policy used to select actionsis the same as the policy used to condition the expectation inthe definition of the value function (Eq. 1). Alternatively, wemight want to make off-policy updates, where the actions arechosen according to some behavior policy b, different fromthe target policy π used in Eq. 1. If we view value estima-tion as estimating the expected return, this off-policy settingcorresponds to estimating an expectation conditioned on onedistribution with samples collected under another. TD canbe extended to make off-policy updates by using importancesampling ratios ρt

def= π(At|St)

b(At|St) ≥ 0. The resulting algorithmis a minor modification of TD, wt+1 ← wt + αtρtδtxt,where δt is defined in Eq. 2.

Off-policy TD can diverge with function approximation,but fortunately there are several TD-based algorithms that

Page 3: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

are convergent. When TD learning converges, it convergesto the TD fixed point: the weight vector where E[δtxt] =0. Interestingly, TD does not perform gradient descent onany objective to reach the TD fixed point. So, one wayto achieve convergence is to perform gradient descent onan objective whose minimum corresponds to the TD-fixedpoint. Gradient TD methods do exactly this on the MeanSquared Projected Bellman Error (MSPBE) (see Eq. 7).

There are several ways to approximate and simplify the gra-dient of MSPBE, each resulting in a different algorithm. Thetwo most well-known approaches are TD with Corrections(TDC) and Gradient TD (GTD2). Both these require doublethe computation and storage of TD, and employ a secondset of learned weights h ∈ Rn with a different stepsizeparameter ηαt, where η is a tunable constant. The updatesfor the TDC algorithm otherwise are similar to TD:

wt+1 ← wt + αtρtδtxt − αtρtγ(h>t xt)xt+1

ht+1 ← ht + ηαt[ρtδt − (h>t xt)

]xt. (3)

The GTD2 algorithm uses the same update for ht, but theupdate to the primary weights is different:

wt+1 ← wt + αtρt(xt − γxt+1)(h>

t xt). (4)

The Gradient TD algorithms are not widely used in practiceand are considered difficult to use. In particular, for TDC,the second stepsize has a big impact on performance (White& White, 2016), and the theory suggests that η > 1 isnecessary to guarantee convergence (Sutton et al., 2009).

Attempts to improve Gradient TD methods has largely comefrom rederiving GTD2 using a saddlepoint formulation ofthe MSPBE (Mahadevan et al., 2014). This formulationenables us to view GTD2 as a one-time scale algorithm witha single set of weights [w,h] using a single global stepsizeparameter. In addition, saddlepoint GTD2 can be combinedwith acceleration techniques like Mirror Prox (Mahadevanet al., 2014) and stochastic variance reduction methods suchas SAGA and SVRG (Du et al., 2017). Unfortunately, Mir-ror Prox has never been shown to improve performanceover vanilla GTD2 (White & White, 2016; Ghiassian et al.,2018). Current variance reduction methods like SAGA areonly applicable in the offline setting, and extension to theonline setting would require new methods (Du et al., 2017).In Appendix B we include comparisons of off-policy predic-tion algorithms in the batch setting, including recent KernelResidual Gradient methods (Feng et al., 2019). These exper-iments suggest that accelerations do not change the relativeranking of the algorithms in the batch setting.

TD is widely considered more sample efficient than all themethods discussed above. A less well-known family ofalgorithms, called Hybrid methods (Maei, 2011; Hackman,2012; White & White, 2016), were designed to exploit the

sample efficiency of TD when data is generated on-policy—they reduce to TD in the on-policy setting—and use gradientcorrections, like TDC, when the data is off-policy. Thesemethods provide some of the ease-of-use benefits of TD,but unfortunately do not enjoy the same level of stability asthe Gradient TD methods: for instance, HTD can divergeon Baird’s counterexample (White & White, 2016).

3. TD with Regularized CorrectionsIn this section we develop a new algorithm, called TD withRegularized Corrections (TDRC). The idea is very simple:to regularize the update to the secondary parameters h. Theinspiration for the algorithm comes from behavior observedin experiments (see Section 4). Consistently, we find thatTDC outperforms—or is comparable to—GTD2 in terms ofoptimizing the MSPBE; as we reaffirm in our experiments.These results match previous experiments comparing thesetwo algorithms (White & White, 2016; Ghiassian et al.,2018). Previous results suggested that TDC could matchTD (White & White, 2016); but, as we highlight in Section4, this is only when the second stepsize is set so small thatTDC is effectively behaving like TD. This behavior is unsat-isfactory because to have guaranteed convergence—e.g. onBaird’s Counterexample—the second stepsize needs to belarge. Further, it is somewhat surprising that attempting toobtain an estimate of the gradient of the MSPBE, as doneby TDC, can perform so much more poorly than TD.

Notice that the h update is simply a linear regression updatefor estimating the (changing) target δt conditioned on xt,for both GTD2 and TDC. As w converges, δt approacheszero, and consequently h goes to 0 as well. But, a linear re-gression estimate of E[δt|St = s] is not necessarily the bestchoice. In fact, using ridge regression—`2 regularization—can provide a better bias-variance trade-off: it can signif-icantly reduce variance without incurring too much bias.This is in particular true for h, where asymptotically h = 0and so the bias disappears.

This highlights a potential reason that TD frequently outper-forms TDC and GTD2 in experiments: the variance of h. IfTD already performs well, it is better to simply use the zerovariance but biased estimate ht = 0. Adding `2 regulariza-tion with parameter β, i.e. β‖h‖22, provides a way to movebetween TD and TDC. For a very large β, h will be pushedclose to zero and the update to w will be lower varianceand more similar to the TD update. On the other hand, forβ = 0, the update reduces to TDC and the estimator h willbe an unbiased estimator with higher variance.

The resulting update equations for TDRC are

wt+1 ← wt + αρtδtxt − αρtγ(h>

t x)xt+1 (5)

ht+1 ← ht + α[ρtδt − (h

>

t xt)]xt − αβht. (6)

Page 4: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

The update to w is the same as TDC, but the update to hnow has the additional term αβht which corresponds to thegradient of the `2 regularizer. The updates only have a singleshared stepsize, α, rather than a separate stepsize for thesecondary weights h. We make this choice precisely for ourmotivated reason upfront: for ease-of-use. Further, we findempirically that this choice is effective, and that the reasonsfor TDC’s sensitivity to the second stepsize are mainly dueto the fact that a small second stepsize enables TDC tobehave like TD (see Section 4.2). Because TDRC has thisbehavior by design, a shared stepsize is more effective.

While there are many approaches to reduce the variance ofthe estimator, h, we use an `2 regularizer because (1) usingthe `2 regularizer ensures the set of solutions for TDRCmatch TD; (2) the resulting update is asymptotically un-biased, because it biases towards the known asymptoticsolution of h; and (3) the strongly convex `2 regularizerimproves the convergence rate. TDC convergence proofsimpose conditions on the size of the stepsize for h to ensurethat it converges more quickly than the “slow-learner” w,and so increasing convergence rate for h should make it eas-ier to satisfy this condition. Additionally, the `2 regularizerbiases the estimator h towards h = 0, the known optimumof the learning system as w converges. This means that thebias imposed on h disappears asymptotically, changing onlythe transient trajectory (we prove this in Theorem 3.2).

As a final remark, we motivate that TDRC should not requirea second stepsize, but have introduced a new parameter (β)to obtain this property. The idea, however, is that TDRCshould be relatively insensitive to β. The choice of β sweepsbetween two reasonable algorithms: TD and TDC. If we arealready comfortable using TD, then it should be acceptableto use TDRC with a larger β. A smaller β will still result ina sound algorithm, though its performance may suffer dueto the variance of the updates in h. In our experiments, wein fact find that TDRC performs well for a wide range ofβ, and that our default choice of β = 1.0 works reasonablyacross all the problems that we tested.

3.1. Theoretically Characterizing the TDRC Update

The MSPBE (Sutton et al., 2009) is defined as

MSPBE(wt)def= E[δtxt]

>E[xtx>t

]−1 E[δtxt] (7)

= (−Aw + b)>C−1(−Aw + b)

where E[δtxt] = b−Awt for

Cdef= E

[xx>

], A

def= E

[x(x− γx′)>

], b

def= E[Rx] .

The TD fixed point corresponds to E[δtxt] = 0 and so to thesolution to the system Awt = b. The expectation is takenwith respect to the target policy π, unless stated otherwise.

The expected update for TD corresponds to E[δtxt] = b−Awt. The expected update for w in TDC corresponds tothe gradient of the MSPBE,

−1

2∇MSPBE(wt) = A>C−1(b−Awt).

Both TDC and GTD2 estimate hdef= C−1(b − Awt) =

E[xtx>t

]−1 E[δtxt], to get the least squares estimateh>xt ≈ E[δt|xt] for targets δt. TDC rearranges terms,to sample this gradient differently than GTD2; for a givenh, both have the same expected update for w: A>h.

We can now consider the expected update for TDRC. Solv-ing for the `2 regularized problem with target δt, weget (E

[xtx>t

]+ βI)h = E[δtxt] which implies hβ =

C−1β (b−Awt) for Cβdef= C+βI. To get a similar form to

TDC, we consider the modified expected update A>βhβ for

Aβdef= A+βI. We can get the TDRC update by rearranging

this expected update, similarly to how TDC is derived

A>β hβ = (E[(x− γx′)x>

]+ βI)hβ

=(E[xx>

]+ βI− γE

[x′x>

])C−1β E[δtxt]

=(E[xx>

]+ βI

)C−1β E[δtxt]− γE

[x′x>

]C−1β E[δtxt]

= E[δtxt]− γE[x′x>

]hβ

This update equation for the primary weights looks pre-cisely like the update in TDC, except that our h is estimateddifferently. Despite this difference, we show in Theorem3.1 (in Appendix G) that the set of TDRC solutions w toA>β hβ = 0 includes the TD fixed point, and this set isexactly equivalent if Aβ is full rank.

In the following theorem (proof in Appendix F) we directlycompare convergence of TDRC to TDC. Though the TDRCupdates are no longer gradients, we maintain the conver-gence properties of TDC. This theorem extends the TDCconvergence result to allow for β > 0, where TDC corre-sponds to TDRC with β = 0.

Theorem 3.1 (Convergence of TDRC) Consider theTDRC update, with a TDC like stepsize multiplier η ≥ 0:

ht+1 = ht +ηαt

[ρtδt − h>t xt

]xt−ηαtβ ht, (8)

wt+1 = wt +αtρtδt xt−αtρtγ(h>t xt)xt+1, (9)

with stepsizes αt ∈ (0, 1], satisfying∑∞t=0 αt = ∞ and∑∞

t=0 α2t < ∞. Assume that (xt, Rt,xt+1, ρt) is an i.i.d.

sequence with uniformly bounded second moments for statesand rewards, A+β I and C are non-singular, and thatthe standard coverage assumption (Sutton & Barto, 2018)holds, i.e. b(A|S) > 0 ∀S,A where π(A|S) > 0. Thenwt converges with probability one to the TD fixed point ifeither of the following are satisfied:

Page 5: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

(i) A is positive definite, or

(ii) β < −λmax(H−1 AA>) and η > −λmin(C−1 H), with

Hdef= A+A>

2 . Note that when A is not positive definite,−λmax(H

−1 AA>) and −λmin(C−1 H) are guaranteed tobe positive real numbers.

We can extend this result to allow for singular C, whichwas not possible for TDC. The set of conditions on η andβ, however, are more complex. We include this result inAppendix ??, with conditions given in Eq. ??.

Theorem 3.2 shows that TDRC maintains convergence whenTD is convergent: the case when A is positive definite.Otherwise, TDRC converges under more general settingsthan TDC, because it has the same conditions on η as givenby Maei (2011) but allows for β > 0. The upper bound onβ makes sense, since as β → ∞, TDRC approaches TD.Examining the proof, it is likely that the conditions on ηcould actually be relaxed (see Eq. C3).

One advantage of TDRC is that the matrix Cβ = C+β Iis non-singular by construction. This raises the question:could we have simply changed the MSPBE objective touse Cβ and derived the corresponding TDC-like algorithm?This is easier than TDRC, as the proof of convergence forthe resulting algorithm trivially extends the proof from Maei(2011), as the change to the objective function is minimal.We derive corresponding TDC-like update and demonstratethat it performs notably worse than TDRC in Appendix A.

4. Experiments in the Prediction SettingWe first establish the performance of TDRC across sev-eral small linear prediction tasks where we carefully sweephyper-parameters, analyze sensitivity, and average overmany runs. The goal is to understand if TDRC has similarperformance to TD, with similar parameter sensitivity, butavoids divergence. Before running TDRC, we set β = 1.0across all the experiments to refrain from tuning this addi-tional parameter.

4.1. Prediction Problems

In the prediction setting, we investigate three different prob-lems with variations in feature representations, target andbehavior policies. We choose problems that have been usedin prior work empirically investigating TD methods. Thefirst problem, Boyan’s chain (Boyan, 2002), is a 13 stateMarkov chain where each state is represented by a compactfeature representation. This encoding causes inappropriategeneralization during learning, but vπ can be representedperfectly with the given features.

Code for all experiments is available at:https://github.com/rlai-lab/Regularized-GradientTD

The second problem is Baird’s (1995) well-known star coun-terexample. In this MDP, the target and behavior policy arevery different resulting in large importance sampling correc-tions. Baird’s Counterexample has been used extensively todemonstrate the soundness of Gradient TD algorithms, soprovides a useful testbed to demonstrate that TDRC doesnot sacrifice soundness for ease-of-use.

Finally, we include a five state random walk MDP. We usethree different feature representations: tabular (unit basisvectors), inverted, and dependent features. This last problemwas chosen so that we could exactly mirror the experimentsused in prior work benchmarking TDC, GTD2, and TD(Sutton et al., 2009). Like Hackman (2012), we used an off-policy variant of the problem. The behavior policy choosesthe left and right action with equal probability, and the targetpolicy chooses the right action 60% of the time. Figure 11in the appendix summarizes all three problems.

We report the total RMSPBE over 3000 steps, measuredon each time step, averaged over 200 independent runs.The learning algorithms under study have tunable meta-parameters that can dramatically impact the efficiency oflearning. We extensively sweep the values of these meta-parameters (as described in Appendix E), and report bothsummary performance and the sensitivity of each method toits meta-parameters. For all results reported in the predic-tion setting, we use the Adagrad (Duchi, Hazan & Singer,2011) algorithm to adapt a vector of stepsizes for each algo-rithm. Additional results for constant scalar stepsizes andADAM vector stepsizes can be found in Appendix B andAppendix D; the conclusions are similar.

4.2. Overall Performance

We first report performance for both the best stepsize as wellas provide the parameter sensitivity plots in Figure 1. Inthe bar plot, we compactly summarize relative performanceto TDRC. TDRC performs well across problems, whileevery other method has at least one setting where it doesnoticeably worse than TDRC. GTD2 generally learns moreslowly than other methods. This result is unsurprising, asit relies so heavily on h for learning w: wt+1 ← wt +

α(xt−γxt+1)h>

t xt. In the beginning, when h is inaccurate,the updates for w are poor. TDC generally learns muchfaster. In Boyan’s chain, however, TDC seems to suffer fromvariance in h. The features in this environment cause biggerchanges in h than in the other environments. TDRC, onthe other hand, which regularizes h, significantly improveslearning in Boyan’s chain. TD and HTD perform very wellacross all problems except for Baird’s. Finally, Vtrace—which uses a TD update with importance sampling ratiosclipped at 1—performs slightly worse than TD due to theintroduced bias, but does not mitigate divergence issues dueto off-policy learning in Baird’s.

Page 6: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Relative

RMSPBE

Tabular Inverted Dependent Boyan Baird

Vtrace

TDRCTD

CGTD2

HTD Vt

race

TDRC

TDCG

TD2

TDHTD

Vtrace

TDRC

TDCG

TD2

HTD TD

RC

TDC

GTD2

TDGTD2

TDRC

TDC

0.51.01.5

2.02.5

3.0

TD TD

RMSPBE

5 4 3 2 1 0 4 3 2 1 0 6 4 2 0 6 4 2 0 5 4 3 2 1 00.04

0.06

0.08

0.10

0.02

0.04

0.06

0.08

0.10

0

1

2

3

0

1

2

3

4

5

0.04

0.06

0.08

0.10

GTD2

TDC Vtrace

TDRC

HTD

TD

TDC

GTD2

TDRC

VtraceTDC

GTD2

TDRC

TD

HTD

Vtrace

GTD2

TDC

TD

TDRC

HTD

TD

HTDTDRC

Vtrace

TDC

GTD2

α = 2−x α = 2−x α = 2−x α = 2−x α = 2−x

Figure 1. Top: The normalized average area under the RMSPBE learning curve for each method on each problem. Each bar is normalizedby TDRC’s performance so that each problem can be shown in the same range. All results are averaged over 200 independent runswith standard error bars shown at the top of each rectangle, though most are vanishingly small. TD and VTrace both diverge on Baird’sCounterexample, which is represented by the bars going off the top of the plot. HTD’s bar is also off the plot due to its oscillating behavior.Bottom: Stepsize sensitivity measured using average area under the RMSPBE learning curve for each method on each problem. HTD andVTrace are not shown in Boyan’s Chain because they reduce to TD for on-policy problems. Values for bar graphs are given in Table 2.

RMSPBEGTD2TDC

TDRC

HTD

TD

TDC

GTD2

TDRC

TDC

GTD2 TDGTD2

TDC

TD

TDRC

HTD

TD

HTD

TDRC

TDC

GTD2

-5 -3 -1 1 3 50.04

0.06

0.08

0.10

0.04

0.06

0.08

0.10

0.02

0.04

0.06

0.08

0.00

0.50

0.75

0.25

1.00

1.25

0

1

2

3

4

-5 -3 -1 1 3 5 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5

TDRCHTD

Tabular Inverted Dependent Boyan Baird

η = 2x η = 2x η = 2x η = 2x η = 2x

Figure 2. Sensitivity to the second stepsize, for changing parameter η. All methods use Adagrad. All methods are free to choose any valueof α for each η. Methods that do not have a second stepsize are shown as a flat line. Values swept are η ∈ 2−6, 2−5, . . . , 25, 26.

The results reported here for TDC do not match previousresults which indicate performance generally as good as TD(White & White, 2016). The reason for this discrepancy isthat previous results carefully tuned the second stepsize ηαfor TDC. The need to tune η is part of the difficulty in usingTDC. To better understand the role it is playing here, weinclude an additional result where we sweep η as well asα for TDC; for completeness, we also include this sweepfor GTD2 and HTD. We sweep η ∈ 2−6, 2−5, . . . , 25, 26.This allows for ηα that is very near zero as well as ηα muchlarger than α. The theory for TDC suggests η should belarger than 1. The results in Figure 2, however, demonstratethat TDC almost always prefers the smallest η; but for verysmall η TDC is effectively a TD update. By picking a smallη, TDC essentially keeps h near zero—its initialization—and so removes the gradient correction term. TDC wastherefore able to match TD by simply tuning a parameterso that it effectively was TD. Unfortunately, this is not ageneral strategy, for instance in Baird’s, TDC picks η ≥ 1and small η perform poorly.

4.3. Sensitivity to β

So far we have only used TDRC with a regularization pa-rameter β = 1. This choice was both to avoid over-tuningour method, as well as to show that an intuitive default value

could be effective across settings. Intuitively, TDRC shouldnot be sensitive to β, as both TDC (β = 0) and TD (largeβ) generally perform reasonably. Picking a β > 0 shouldenable TDRC to learn faster like TD—by providing a lowervariance correction—as long as it’s not too large, to ensurewe avoid the divergence issues of TD.

We investigate this intuition by looking at performanceacross a range of β ∈ 0.1 ∗ 20, 21, . . . , 25, 26. For β = 0,we have TDC. Ideally, performance should quickly improvefor any non-negligible β, with a large flat region of goodperformance in the parameter sensitivity plots for a widerange of β. This is generally what we observe in Figure 3.For even very small β, TDRC noticeably improves perfor-mance over TDC, getting halfway between TDC and TD(Random Walk with Tabular or Dependent features) or insome cases immediately obtaining the good performance ofTD (Random Walk with Inverted Features, Boyan’s chainand Baird’s). Further, in these three cases, it even performsbetter or comparably to both TDC and TD for all tested β.Notably, these are the settings with more complex featurerepresentations, suggesting that the regularization parame-ter helps TDRC learn an h that is less affected by harmfulaliasing in the feature representation. Finally, the resultsalso show that β = 1.0 was in fact not optimal, and wecould have obtained even better results in the previous sec-

Page 7: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

0.04

0.06

TDRC

0.05

TDC

TD0.040

0.050

TDC

TD

TDRCTDRC

TD

TDC

TD

TDC

TDRC

0.2

0.6

0.4

0.3

0.5

-3 2-1-2 0 1 -3 2-1-2 0 1 -3 2-1-2 0 1 -3 2-1-2 0 1

TD

TDRC

4.0

-3 2-1-2 0 10.0

1.0

2.0

3.0

Tabular Inverted Dependent Boyan Baird

RMSPBETDC

0.022

0.027

0.032

0.045

0.055

β = 2x β = 2x β = 2x β = 2x β = 2x

Figure 3. Sensitivity to the regularization parameter, β. TD and TDC are shown as dotted baselines, demonstrating extreme values of β;β = 0 represented by TDC and β →∞ represented by TD. This experiment demonstrates TDRC’s notable insensitivity to β. Its similarrange of values across problems, including Baird’s counterexample, motivates that β can be chosen easily and is not heavily problemdependent. Values swept are: β ∈ 0.1 ∗ 20, 21, . . . , 25, 26.

tion, typically with a larger β. These improvements, though,were relatively marginal over the choice of β = 1.0.

Naturally, the scale of β should be dependent on the magni-tude of the rewards, because in TDRC the gradient correc-tion term is attempting to estimate the expected TD error.One answer is to simply employ adaptive target normaliza-tion, such as Pop-Art (van Hasselt et al., 2016), and keepβ equal to one. We found TDRC with β = 1 performedat least as well as TD in on-policy chain domains across alarge range of reward scales (see Appendix ??).

5. Experiments in the Control SettingLike TD, TDRC was developed for prediction, under linearfunction approximation. Again like TD, there are natural—though in some cases heuristic—extensions to the controlsetting and to non-linear function approximation. In this sec-tion, we investigate if TDRC can provide similar improve-ments in the control setting. We first investigate TDRC incontrol with linear function approximation, where the exten-sion is more straightforward. We then provide a heuristicstrategy to use TDRC—and TDC—with non-linear functionapproximation. We demonstrate, for the first time, that Gra-dient TD methods can outperform Q-learning when usingneural networks, in two classic control domains and twovisual games.

5.1. Extending TDRC to Control

Before presenting the control experiments, we describehow to extend TDRC to control, and to non-linear func-tion approximation. The extension to non-linear functionapproximation is also applicable in the prediction setting;we therefore begin there. We then discuss the extension toQ-learning which involves estimating action-values for thegreedy policy.

Consider the setting where we estimate v(s) using a neu-ral network. The secondary weights in TDRC are used toobtain an estimate of E[δt|St = s]. Under linear functionapproximation, this expected TD error is estimated using lin-ear regression with `2 regularization: h>xt ≈ E[δt|St = s].

With neural networks, this expected TD error can be esti-mated using an additional head on the network. The tar-get for this second head is still δt, with a squared errorand `2 regularization. One might even expect this estimateof E[δt|St = s] to improve, when using a neural network,rather than a hand-designed basis.

An important nuance is that gradients are not passed back-ward from the error in this second head. This choice is madefor simplicity, and to avoid any issues when balancing thesetwo losses. The correction is secondary, and we want toavoid degrading performance in the value estimates simplyto improve estimates of E[δt|St = s]. It also makes theconnection to TD more clear as β becomes larger, as theupdate to the network is only impacted by w. We have notextensively tested this choice; it remains to be seen if usinggradients from both heads might actually be a better choice.

The next step is to extend the algorithm to action-values. Foran input state s, the network produces an estimate q(s, a)

and a prediction δ(s, a) of E[δt|St = s,At = a] for eachaction. The weights ht+1,At for the head corresponding toaction At are updated using the features produced by thelast layer xt, with δ(St, At) = h

>

t,Atxt:

ht+1,At ← ht,At + α[δt − h

>

t,Atxt]xt − αβht,At (10)

For the other actions, the secondary weights are not updatedsince we did not get a target δt for them.

The remaining weights wt, which include all the weights inthe network excluding h, are updated using

δt = Rt+1 + γq(St+1, a′)− q(St, At) (11)

wt+1←wt+αδt∇wq(St, At)−αγδ(St, At)∇wq(St+1, a′)

where a′ is the action that the policy we are evaluatingwould take in state St+1. For control, we often select thegreedy policy, and so a′ = arg maxa q(St+1, a) and δt =Rt+1 + γmaxa q(St+1, a) − q(St, At) as in Q-learning.This action a′ may differ from the (exploratory) action At+1

that is actually executed, and so this estimation is off-policy.There are no importance sampling ratios because we areestimating action-values.

Page 8: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

We call this final algorithm QRC: Q-learning with Regu-larized Corrections. The secondary weights in QRC areinitialized to 0, to maintain the similarity to TD. We canobtain, as a special case, a control algorithm based on TDC,which we call QC. If we set β = 0 in Eq. 11, we obtain QC.

We conclude this section by highlighting that there is analternative route to use TDRC, as is, for control: by usingTDRC as a critic within Actor-Critic. We provide the updateequations in Appendix ??.

5.2. Control Problems

We first test the algorithms in a well-understood setting, inwhich we know Q-learning is effective: Mountain Car with atile-coding representation. We then use neural network func-tion approximation in two classic control environments—Mountain Car and Cart Pole—and two visual environmentsfrom the MinAtar suite (Young & Tian, 2019). For all en-vironments, we fix β = 1.0 for QRC, η = 1.0 for QCand do not use target networks (for experiments with targetnetworks see Appendix ??).

In the two classic control environments, we use 200 runs,an ε-greedy policy with ε = 0.1 and a discount of γ = 0.99.In Mountain Car (Moore, 1990; Sutton, 1996), the goal isto reach the top of a hill, with an underpowered car. Thestate consists of the agent’s position and velocity, with areward of −1 per step until termination, with actions toaccelerate forward, backward or do nothing. In Cart Pole(Barto, Sutton & Anderson, 1983), the goal is to keep a polebalanced as long as possible, by moving a cart left or right.The state consists of the position and velocity of the cart,and the angle and angular velocity of the pole. The reward is+1 per step. An episode ends when the agent fails to balancethe pole or balances the pole for more than 500 consecutivesteps. For non-linear control experimental details on theseenvironments see Appendix ??.

For the two MinAtar environments, Breakout and SpaceInvaders, we use 30 runs, γ = 0.99 and a decayed ε-greedypolicy with ε = 1 decaying linearly to ε = 0.1 over thefirst 100,000 steps. In Breakout, the agent moves a paddleleft and right, to hit a ball into bricks. A reward of +1 isgiven for every brick hit; new rows appear when all the rowsare cleared. The episode ends when the agent misses theball and it drops. In Space Invaders, the agent shoots alienships coming towards it, and dodges their fire. A reward of+1 is given for every alien that is shot. The episode endswhen the spaceship is hit by alien fire or reached by an alienship. These environments are simplified versions from theAtari suite, designed to avoid the need for large networksand make it more feasible to complete more exhaustivecomparison, including using more runs. All methods use anetwork with one convolutional layer, followed by a fullyconnected layer. All experimental settings are identical to

the original MinAtar paper (see Appendix ?? for details).

5.3. Linear Control

We compare TD, TDC and TDRC for control, both withinan Actor-Critic algorithm and with their extensions to Q-learning. In Figure ??, we can see two clear outcomesfrom both control experiments. In both cases, the controlalgorithm based on TDC fails to converge to a reasonablepolicy. The TDRC variants, on the other hand, match theperformance of TD.

Q-learning QRC

QC

0

500

1000

1500

2000

# Steps /

episode

AC-TD

AC-TDRC

AC-TDC

Steps ( )×103Steps ( )×103

Actor-Critic Action-value

0 12.5 25 37.5 505 150 10 20

Figure 4. Numbers of steps to reach goal, averaged over runs, ver-sus number of environment steps, in Mountain Car with tile-codedfeatures. Left: Comparison of actor-critic control algorithms withvarious critics with ADAM optimizer. For actor critic experimen-tal details see Appendix ??. Right: Comparison of state-actionvalue control algorithms with constant stepsizes. Stepsizes wereswept over α ∈ 2−8, 2−7, . . . , 2−2, 2−1 and then scaled by thenumber of active features. We used 16 tilings and 4 × 4 tiles.Results are averaged over 200 independent runs, with shaded errorcorresponding to standard error.

This result might be surprising, since the only differencebetween TDRC and TDC is regularizing h. This small addi-tion, though, seems to play a big role in avoiding this sur-prisingly bad performance of TDC, and potentially explainswhy gradient methods have been dismissed as hard-to-use.When we looked more closely at TDC’s behavior, we foundthat the TDC agent improved its behavior policy quickly.But, the magnitude of the gradient corrections also grewrapidly. This high magnitude gradient correction resultedin a higher magnitude gradient for w, and pushed down thelearning rate for TDC. The constraint on this correction termprovided by TDRC seems to prevent this explosive growth,allowing TDRC to attain comparable performance to theTD-based control agent.

5.4. Non-linear Control

When moving to non-linear function approximation, withneural networks, we find a more nuanced outcome: QC stillsuffers compared to Q-learning and QRC in the classic con-trol environments—though less than before—yet providessubstantial improvements in the two MinAtar environments.

In Figure ??, we find that QC learns more slowly than QRCand Q-learning. Again, QRC brings performance much

Page 9: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

# Steps /

episodeDQN

DQC

DQRC200

400

600

800

1000

5 150 10 20 25

Steps (×103)

QRC

QC

Q-learning

QRC

QC

Steps(×104)

0

100

200

300

400

# Steps /

episode

Q-learning

DQRC

DQN

DQC

200

400

600

800

1000

α = 2−x

QRC

QC

Q-learning

12 10 8 6 4 216 14

2 60 4 8 10

QRC

QC

12 10 8 6 4 2

α = 2−x

0

100

200

300

400

Q-learning

Mountain Car Cart Pole

Figure 5. Performance of Q-learning, QC and QRC on two classiccontrol environments. On top the learning curves are shown and atthe bottom the parameter sensitivity for various stepsizes. Loweris better for Mountain Car (fewer steps to goal) and higher is betterfor Cart Pole (more steps balancing the pole). Results are averagedover 200 runs, with shaded error corresponding to standard error.

closer to Q-learning, when QC is performing notably morepoorly. In Mountain Car, we tested a more highly off-policysetting: 10 replay steps. By using more replay per step,more data from older policies is used, resulting in a more off-policy data distribution. Under such an off-policy setting,we expect Q-learning to suffer, and in fact, we find thatQRC actually performs better than Q-learning. We provideadditional experiments on Mountain Car in Appendix ??.

On the two MinAtar environments, in Figure 5, we obtaina surprising result: QC provides substantial performanceimprovements over Q-learning. QRC with β = 1 is not asperformant as QC in this setting and instead obtains perfor-mance in-between QC and Q-learning. However, QRC withsmaller values of regularization parameter (shown as lighterblue lines) results in the best performance. This outcomehighlights that Gradient TD methods are not only theoreti-cally appealing, but could actually be a better alternative toQ-learning in standard (non-adversarially chosen) problems.It further shows that, though QRC with β = 1.0 generallyprovides a reasonable strategy, substantial improvementscould be obtained with an adaptive method for selecting β.

6. Conclusions and DiscussionIn this work, we introduced a simple modification of theTDC algorithm that achieves performance much closer tothat of TD. Our algorithm uses a single stepsize like TD, andbehaves like TD when TD performs well but also prevents di-vergence under off-policy sampling. TDRC is built on TDC,and, as we prove, inherits its soundness guarantees. In smalllinear prediction problems TDRC performs best overall andexhibits low sensitivity to its regularization parameter. Incontrol experiments, with extensions to non-linear function

Total discounted

reward averaged

over 30 runs

QRC

Q-learning

QC

Moving average

of returns over

100 episodesQ-learning

QRC

11

Breakout Space Invaders

Q-learning

QRC

Q-learning

QC

α = 2−x910

0.10.01

0.0010.0001

QRC

0.1

0.01

0.0010.0001

0.01

QC

0.0010.0001

0.1

0.01

0.1

0.001

0.0001

0.00

0.05

0.10

0.15

0.20

0.1

0.2

0.3

0.0

0.4

0.00

0.05

0.10

0.15

0.1

0.2

0.3

0.0910 8

α = 2−x7 6 5 12 8

0 1 2 3 4 51 3 50 2 4

Steps ( )×106 Steps ( )×106

Figure 6. Performance of Q-learning, QC, and QRC in the twoMinAtar environments. The learning curves in the top row depictthe average return over time for the best performing stepsize foreach agent. The stepsize sensitivity plots in the bottom row depictthe total discounted reward achieved with several stepsize values.Higher is better. Results are averaged over 30 independent runs,with shaded error corresponding to standard error. Light bluelines show the performance of QRC with smaller regularizationparameters, β < 1.

approximation, we find that the resulting algorithm, QRC,performs as well as Q-learning and in some cases notablybetter. This constitutes the first demonstration of GradientTD methods outperforming Q-learning, and suggests thissimple modification to the standard Q-learning update—togive QRC—could provide a more general purpose algo-rithm.

An important next step is to better understand the conditionson the regularization parameter β and whether we can trulyremove the second stepsize η. The current theorem does notremove conditions on η; in fact, it has the same conditionsas TDC. We hypothesize that β should make h convergemore quickly, and so remove the need for the stepsize forthe secondary weights to be bigger. Further, the conditionson η and β both depend on domain specific quantities thatare generally difficult to compute. In the small predictionproblems, we were easily able to confirm that our choices ofmeta-parameter met the theoretical conditions, however forthe larger control problems this remains an open question.In general, developing tight conditions on η and β wouldhelp facilitate comfort in using TDRC.

Another important next step is to thoroughly investigate ifthese empirical results hold in a broader range of environ-ments and settings. The results in this work suggest thatTDRC could potentially be a replacement for the widelyused TD algorithms. It is only a small modification to anexisting TD implementation, and so would not be difficult toadopt. But, to make such a bold claim, much more evidenceis needed, particularly because TD has been shown to be sosuccessful for many years.

Page 10: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

AcknowledgmentsThis work was funded by NSERC and CIFAR, particularlythrough funding the Alberta Machine Intelligence Institute(Amii) and the CCAI Chair program. The authors alsogratefully acknowledge funding from JPMorgan Chase &Co. and Google DeepMind. We would like to thank CsabaSzepesvari, Anna Harutyunyan, and the anonymous review-ers for useful feedback. We also thank Banafsheh Rafiee,Andrew Jacobsen, and Alan Chan for helpful discussionsduring the course of this project.

ReferencesBaird, L. C. (1995). Residual algorithms: Reinforcement

learning with function approximation. In InternationalConference on Machine Learning, pp. 30–37. MorganKaufmann, San Francisco.

Barto, A. G., Sutton, R. S., Anderson, , C. W. (1983). Neu-ronlike adaptive elements that can solve difficult learningcontrol problems. IEEE transactions on systems, man,and cybernetics, 5, 834-846.

Bellemare, M. G., Naddaf, Y., Veness, J., Bowling, M.(2013). The arcade learning environment: An evalu-ation platform for general agents. Journal of ArtificialIntelligence Research, 47, 253-279.

Borkar, V. S., Meyn, S.P. (2000). The O.D.E. Method forConvergence of Stochastic Approximation and Rein-forcement Learning. SIAM J. Control and Optimization.

Boyan, J.A. (2002). Technical Update: Least-Squares Tem-poral Difference Learning. Machine Learning.

Dai, B., Albert, S., Lihong, L., Lin, X., Niao, H., Zhen,L., Jianshu, C., Le, S. SBEED: Convergent reinforce-ment learning with nonlinear function approximation.In International Conference on Machine Learning (pp.1125-1134).

Du, S. S., Chen, J., Li, L., Xiao, L., and Zhou, D. (2017)Stochastic variance reduction methods for policy evalua-tion. In International Conference on Machine Learning.

Duchi, J., Hazan, E., Singer, Y. (2011). Adaptive subgradi-ent methods for online learning and stochastic optimiza-tion. Journal of Machine Learning Research 12:2121-2159.

Espeholt, L., Soyer, H., Munos, R., Simonyan, K., Mnih,V., Ward, T., Doron, Y., Firoiu, V., Harley, T., Dunning,I. and Legg, S. (2018) IMPALA: Scalable distributedDeep-RL with importance weighted actor-learner ar-chitectures. In International Conference on Machine

Learning.

Feng, Y., Li, L., Liu, Q. (2019). A kernel loss for solving thebellman equation. In Advances in Neural InformationProcessing Systems (pp. 15430-15441).

Ghiassian, S., Patterson, A., White, M., Sutton, R.S., White, A. (2018). Online off-policy prediction.ArXiv:1811.02597.

Glorot, X., Bengio, Y. (2010). Understanding the difficultyof training deep feedforward neural networks. In In-ternational Conference on Artificial Intelligence andStatistics (pp. 249-256).

Hackman, L. (2012). Faster Gradient TD Algorithms. M.Sc.thesis, University of Alberta, Edmonton.

Juditsky, A., Nemirovski, A. (2011). Optimization for Ma-chine Learning.

Kingma, D. P., Ba, J. (2014). Adam: A method for stochas-tic optimization. arXiv preprint arXiv:1412.6980.

Liu B, Liu J, Ghavamzadeh M, Mahadevan S, Petrik M(2015). Finite-Sample Analysis of Proximal GradientTD Algorithms. In International Conference on Uncer-tainty in Artificial Intelligence, pp. 504-513.

Liu B, Liu J, Ghavamzadeh M, Mahadevan S, Petrik M(2016). Proximal Gradient Temporal Difference Learn-ing Algorithms. In International Joint Conference onArtificial Intelligence, pp. 4195-4199.

Mahadevan, S., Liu, B., Thomas, P., Dabney, W., Giguere,S., Jacek, N., Gemp, I., Liu, J. (2014). Proximal rein-forcement learning: A new theory of sequential decisionmaking in primal-dual spaces. ArXiv:1405.6757.

Mahmood, A. R., Yu, H., Sutton, R. S. (2017). Multi-stepoff-policy learning without importance sampling ratios.ArXiv:1702.03006.

Maei, H. R. (2011). Gradient temporal-difference learningalgorithms. Ph.D. thesis, University of Alberta, Edmon-ton.

Mnih, V., Badia, A. P., Mirza, M., Graves, A., Lillicrap, T.,Harley, T., Silver, D., Kavukcuoglu, K. (2016). Asyn-chronous methods for deep reinforcement learning. InInternational conference on machine learning (pp. 1928-1937).

Moore, A. W. (1990). Efficient memory-based learning forrobot control. Ph.D. theis, University of Cambridge.

Page 11: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Munos, R., Stepleton, T., Harutyunyan, A., Bellemare, M.(2016). Safe and efficient off-policy reinforcement learn-ing. In Advances in Neural Information Processing Sys-tems 29, pp. 1046–1054.

Precup, D., Sutton, R. S., Dasgupta, S. (2001). Off-policytemporal-difference learning with function approxima-tion. In Proceedings of the 18th International Confer-ence on Machine Learning, pp. 417–424.

Reddi, S. J., Kale, S., Kumar, S. (2019). On the convergenceof adam and beyond. ArXiv:1904.09237.

Schaul, T., Quan, J., Antonoglou, I., Silver, D. (2016). Pri-oritized experience replay. In International Conferenceon Learning Representations.

Sutton, R. S. (1996). Generalization in reinforcement learn-ing: Successful examples using sparse coarse coding. InAdvances in Neural Information Processing Systems 8(NIPS 1995), pp. 1038–1044. MIT Press, Cambridge,MA.

Sutton, R. S., Barto, A. G. (2018). Reinforcement Learning:An Introduction, Second Edition. MIT Press.

Sutton, R. S., Maei, H. R., Precup, D., Bhatnagar, S., Silver,D., Szepesvari, Cs., Wiewiora, E. (2009). Fast gradient-descent methods for temporal-difference learning withlinear function approximation. In International Confer-ence on Machine Learning, pp. 993–1000, ACM.

Sutton, R. S., Mahmood A. R., and White M. (2016) An em-phatic approach to the problem of off-policy temporal-difference learning. The Journal of Machine LearningResearch.

Touati, A., Bacon, P. L., Precup, D., Vincent, P. (2018).Convergent tree-backup and retrace with function ap-proximation. ArXiv:1705.09322.

van Hasselt, H. P., Guez, A., Hessel, M., Mnih, V., Silver,D. (2016). Learning values across many orders of mag-nitude. In Advances in Neural Information ProcessingSystems (pp. 4287-4295).

van Hasselt, H., Doron, Y., Strub, F., Hessel, M., Sonnerat,N., Modayil, J. (2018). Deep Reinforcement Learningand the Deadly Triad. ArXiv:1812.02648

White, A., White, M. (2016). Investigating Practical LinearTemporal Difference Learning. In International Confer-ence on Autonomous Agents & Multiagent Systems.

Young, K., Tian, T. (2019). MinAtar: An Atari-InspiredTestbed for Thorough and Reproducible ReinforcementLearning Experiments. ArXiv:1903.03176.

Page 12: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

A. Results in the Batch SettingThe proofs of convergence for many of the methods requireindependent samples for the updates. This condition isnot generally met in the fully online learning setting thatwe consider throughout the rest of the paper. In Figure 6we show results for all methods in the fully offline batchsetting, demonstrating that—on the small problems that weconsider—the conclusions do not change when transferringfrom the batch setting to the online setting. We include twoadditional methods in the batch setting, the Kernel ResidualGradient methods (Feng, Li & Liu, 2019), which do nothave a clear fully online implementation.

We create a new batch dataset for each of 500 independentruns by getting 100k samples from the state distributioninduced by the behavior policy, then sampling from thetransition kernel for each of these states. We then performmini-batch updates by sampling 8 independent transitionsfrom this dataset. Each algorithm makes n updates forn ∈ [1, 2, 4, 8, . . . , 8192], choosing the stepsize which mini-mizes the area under the RMSPBE learning for each n. Thiseffectively shows the best performance of each algorithm ifit was given a budget of n updates, allowing us to make com-parisons across several different timescales. The constantstepsizes swept are α ∈ 2−8, 2−7, . . . , 20.In Figure 6, we demonstrate that GTD2 and the Kernel-RG methods generally perform poorly across these setof domains. We additionally show that TDC, TD, andTDRC are often indistinguishable in the batch setting—except Boyan’s Chain where TDC still performs inexpli-cably poorly—suggesting that perhaps TDRC’s gain in per-formance of TDC is due to the correlated sampling inducedby online learning. We finally show that TDC++, which isTDC with regularized C, generally performs comparably toGTD2.

A.1. Relationship to Residual Gradients

The Residual Gradient (RG) family of algorithms provide analternative gradient-based strategy for performing temporaldifference learning. The RG methods minimize the MeanSquared Bellman Error (MSBE), while the Gradient TD fam-ily of algorithms minimize a particular form of the MSBE,the Mean Squared Projected Bellman Error (MSPBE). TheRG family of methods generally suffer from difficulty inobtaining independent samples from the environment, lead-ing towards stochastic optimization algorithms which finda biased solution (Sutton & Barto, 2018). However, veryrecent work has generalized the MSBE and proposed analgorithmic strategy to perform unbiased stochastic updates(Feng, Li & Liu, 2019; Dai et al., 2018). We compare to theapproach in Feng, Li, and Liu (2019) below.

A.2. Derivation of the TDC++ Update Equations

In this section, we derive the update equations for TDC++,i.e. TDC with the regularized Cβ matrix. Consider theMSPBE objective (see Eq. 7) but with a regularized Cβ :

MSPBE++(wt)def= E[δtxt]

>(E[xtx>t

]−1+ β I

)E[δtxt]

= (−Aw + b)>C−1β (−Aw + b).

The gradient of this objective is − 12∇wMSPBE++(wt) =

A>C−1β (b−Awt) = E[δtxt] − γE[x′x>

]hβ − β hβ .

Using this gradient and the same update for ht+1 as inTDRC, we obtain the update equations for TDC++ (with anadditional η in the stepsize for h):

ht+1 ← ht + ηα[δt − (h

>

t xt)]xt − ηαβht

wt+1 ← wt + αδtxt − αγ(h>

t x)xt+1 − αβht.

A.3. Convergence of TDC++

It is straightforward to show that TDC++ converges to theTD fixed point under very similar conditions as TDC (Maei,2011). We show the key steps here (for details see Maei(2011) or Appendix F). The G matrix for TDC++ is G =[ −ηCβ −ηAA>−Cβ −A

]. If we can show that the real parts of

all the eigenvalues of G are negative, then the algorithmwould converge. First note that for an eigenvalue λ ∈ C ofG, det(G−λ I) = det(λ(Cβ +λ I)+A(ηA>+λ I)) = 0.Then for some non–zero vector z ∈ C, z∗(λ(Cβ +λ I) +

A(ηA>+λ I)) z = 0. Upon simplifying this, we obtainthe following quadratic equation in λ:

‖ z ‖2λ2 + (z∗(ηCβ +A) z)λ+ η‖Az ‖2 = 0.

If λ1 and λ2 are two solutions of this equation, then

λ1λ2 = η‖Az ‖2‖ z ‖2 , λ1 + λ2 = − (z∗(ηCβ +A) z)

‖ z ‖2 .

Since, λ1λ2 > 0 and real, the real parts of both λ1and λ2 have the same sign. Thus, Re(λ1 + λ2) < 0would imply that each of Re(λ1) < 0 and Re(λ2) <0 and we would be done. Assuming Re(λ1 + λ2) =

− (z∗(ηCβ +A) z)∗+(z∗(ηCβ +A) z)2‖ z ‖2 = −z∗(ηCβ +H) z

‖ z ‖2 < 0,

where Hdef= 1

2 (A+A>), leads to the condition

η > −λmin(C−1β H),

for TDC++ to converge.

TDC++ differs from TDRC in that it has an extra term(−αβ ht) in the update for the weight wt+1. Further, un-like TDRC, the convergence of TDC++ doesn’t require anyconditions on β.

Page 13: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Tabular

101 102

Number of Updates

0.05

0.10

0.15

0.20

RMSP

BE

TDCTDGTD2TDRCTDC++GaussianKernelLinearKernel

Inverted

101 102

Number of Updates

0.05

0.10

0.15

0.20

0.25

Dependent

101 102

Number of Updates

0.00

0.05

0.10

0.15

Boyan

101 102

Number of Updates

0.5

1.0

1.5

2.0

2.5

Baird

101 102

Number of Updates

1

2

3

4 TDCGTD2TDRCTDC++GaussianKernelLinearKernel

Figure 7. Sensitivity to the number of update steps for the offline batch setting. Each problem used a dataset of 100k samples sampledfrom the stationary distribution, then mini-batch updates used 8 independent samples from the dataset. On the x-axis we show alog-scale number of updates for each algorithm, on the y-axis we show the area under the RMSPBE learning curve averaged over 500independent runs and 500 independently sampled datasets, with shaded regions showing the standard error over runs. For each numberof update steps shown, we sweep over stepsizes and select the best stepsize for that number of updates; stepsizes were swept fromα ∈ 2−5, 2−4, . . . , 20. For TDRC, we set β = 1. This effectively shows the best performance of each algorithm if it was only given afixed number of updates. GTD2 and the Kernel-RG methods show notably slower convergence than other methods.

Relative

RMSPBE

Tabular Inverted Dependent Boyan Baird

Vtrace

TDRC

TDCG

TD2

HTD Vt

race

TDRC

TDCG

TD2

TDHTD

Vtrace

TDRC

TDCG

TD2

HTD TD

RC

TDCG

TD2

TDGTD2

TDRCTD

C

0.51.0

1.5

2.02.5

3.0

TD TD

Figure 8. Relative performance of methods using the Adam stepsize selection algorithm, compared using the average area under theRMSPBE learning curve. Values swept are: α ∈ 2−8, 2−7, . . . , 2−1, 20 and as before, we set β = 1 for TDRC. On Baird’scounterexample, TD, HTD, and VTrace all exhibit slow learning as well. The actual number for area under the learning curve are shownin Table ??.

B. Incorporating AccelerationsTrue stochastic gradient methods provide the benefit thatthey should be amenable to accelerations for stochasticapproximation, such as momentum, mirror-prox updates(Juditsky & Nemirovski, 2011), and variance reductiontechniques (Du et al., 2017). This is in fact one of thearguments motivating GTD2, and its formulation as a sad-dlepoint method.

We begin investigating how acceleration in the online pre-diction setting impacts the overall performance and relativeordering of the algorithms. Momentum is commonly usedin online deep RL systems, and is a form of acceleration.We compare all the methods using Adam (Kingma & Ba,2014; Reddi, Kale & Kumar, 2019), which includes mo-mentum. Several recently proposed optimizers include mo-mentum and are best viewed as extensions of Adam. Herewe use Adam as there is little evidence in the literature thatthese new variants are better than Adam for online updates.We sweep over values of the meta-parameters in Adam,β1, β2 ∈ 0.9, 0.99, 0.999, and select the values that bestminimize the total RMSPBE separately for each algorithm.

The bar plot in Figure 7 parallels Figure 1, which uses Ada-grad, with similar conclusions. The only notable differenceis that TDC’s performance on Boyan’s chain is much better,though it is still not as good as TD and TDRC. Overall,the use of momentum did not accelerate convergence, with

performance similar to Adagrad. The comparison is notperfect, as Adagrad allows the stepsizes to decrease to zero,which enables the algorithms to converge nicely on these do-mains. Adam does not due to the exponential average in thesquared gradient term. These results, then, mainly providea sanity check that results under an alternative optimizer areconsistent with the previous results.

The majority of accelerations that can be used in policyevaluation are designed for off-line batch updates. Althoughwe are more concerned with online performance, we use thebatch setting in Appendix A as a sanity check to ensure thatnone of the recently proposed accelerated policy evaluationmethods significantly outperform TD, TDC, or TDRC. Inaddition we include Kernel Residual Gradient (Kernel-RG)(Feng, Li & Liu, 2019). Figure 6 shows the performanceof several methods given a fixed budget number of updates.Surprisingly, the Kernel-RG methods show much slowerconvergence across all problems tested.

C. Sensitivity to the Scale of hIn Figure 3 we demonstrate TDRC’s sensitivity to the reg-ularization weight, β, which is responsible for balancingbetween the loss due to the regularizer and the mean-squarederror for h. We motivate empirically that, on a set of smalldomains, the scale of the regularizer does not significantlyaffect the performance of TDRC. However, as the scale of h

Page 14: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Tabular

<latexit sha1_base64="EpwadKl79nuFFVBzWqBryCC0A38=">AAAB7HicbVBNS8NAEJ34WetX1aOXxSJ4KkkV9Fj04rGCaQttKJvtpl262YTdiVBCf4MXD4p49Qd589+4bXPQ1gcDj/dmmJkXplIYdN1vZ219Y3Nru7RT3t3bPzisHB23TJJpxn2WyER3Qmq4FIr7KFDyTqo5jUPJ2+H4bua3n7g2IlGPOEl5ENOhEpFgFK3k90KOtF+pujV3DrJKvIJUoUCzX/nqDRKWxVwhk9SYruemGORUo2CST8u9zPCUsjEd8q6lisbcBPn82Ck5t8qARIm2pZDM1d8TOY2NmcSh7YwpjsyyNxP/87oZRjdBLlSaIVdssSjKJMGEzD4nA6E5QzmxhDIt7K2EjaimDG0+ZRuCt/zyKmnVa95lrf5wVW3cFnGU4BTO4AI8uIYG3EMTfGAg4Ble4c1Rzovz7nwsWtecYuYE/sD5/AHFJI6o</latexit>

Reward magnitude

Inverted

Reward magnitude

Dependent

Reward magnitude

Figure 9. Relationship between TDRC and TD performance across different reward scales for different values of beta. On the x-axiswe show the scale of the rewards for the terminal states of the random walk, on the y-axis we show a range of values of β. Each dotrepresents the number of standard deviations away from TD that TDRC’s performance is across 500 independent runs for that particularvalue of β. For each dot, TDRC and TD choose the stepsize with lowest area under the RMSPBE learning curve; with stepsizes sweptfrom α ∈ 2−5, 2−4, . . . , 20. As the scale of the rewards increases (left to right on the x-axis), the variance of the secondary weights, h,also increases; effectively requiring a larger value of β. This figure demonstrates that TDRC with β = 1 remains relatively insensitive tothe scale of the rewards except in extreme cases when the variance of the rewards from transition to transition is quite large.

varies we likewise expect the scale of β to vary accordingly.

We design a set of small experiments to understand howchanges in the environment cause the scale of h to change,and how that relates to the performance of TDRC across sev-eral values of β. The scale of h changes whenever the sizeof the TD error or scale of the features change. For theseexperiments, we chose to increase the range of the TD errorby making the initial value function V = 0 and manipulat-ing the magnitude of the rewards. We run this experimenton the five state random-walk domain with each of the fea-ture representations used in Section 4, and change only therewards in the terminal states by a multiplicative constant.We compute the mean and standard deviation of TD’s per-formance across 500 independent runs and compute thenumber of standard deviations TDRC’s mean performanceis from TD’s mean performance. We let the reward varyby order of magnitudes, with the multiplicative constanttaking values 10−2, 10−1, . . . , 103. For each scaling, wetest multiple values of β ∈ 2−5, 2−4, . . . , 24 and for eachof these instances we select the best constant stepsize from2−5, 2−4, . . . , 2−1.In Figure 9, we show the range of β for which TDRC’s per-formance is as good, or nearly as good, as TD’s performanceas the magnitude of the rewards increases. As hypothesized,the range of acceptable β decreases as the reward magnitudeincreases; however, the range of β only appreciably shrinksfor a pathologically large deviation between rewards andinitial value function. This demonstrates that, while β isproblem dependent, its range of acceptable values is robustto all but the most pathological of examples across severaldifferent representations.

D. Investigating QC on Mountain CarIn this section we include a deeper preliminary investiga-tion into the performance of QC on the Mountain Car en-vironment with non-linear function approximation. As weobserved in Figure ??, QC performed considerably worsethan either Q-learning and QRC. We hypothesize that thispoor performance is the result of high variance updatesto the value function estimate due to a poor estimate ofE[δt | S = st]. We relax the restrictions on the secondarystepsize, ηα, by using η = 1

2 , allowing QC to become morelike Q-learning and reducing the variance of the updateto the secondary weights. We conclude by investigatingthe effects of prioritization of the replay buffer by drawingsamples according to the squared TD error.

We start by investigating the performance of each algorithmwhen only a single step of replay is used on each environ-mental step. The learning curve in Figure ?? reaffirms thatQRC and Q-learning significantly outperform QC in thissetting. Interestingly, the norm of QC’s secondary set ofweights grows nearly monotonically throughout learningwhile in contrast, QRC’s secondary weights start large atthe beginning of learning and quickly shrink as the valuefunction estimates become more accurate. The bottom rightcurve shows the mean and standard deviation of the max-imum absolute value of q(St, ·) for each step of learning.The variance of QC’s maximum state-action value increasedsignificantly over the maximum observable return in theMountain Car domain—which is represented by a dashedline at 100. These plots in combination suggest that QRC’sadditional constraint on the magnitude of the secondaryweights helps stabilize the learning system when using neu-ral network function approximators.

One plausible explanation for QC’s poor performance is that

Page 15: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

# Steps /

episode

DQRC

DQN

DQC

DQN

DQC

DQRC200

400

600

800

1000

300

500

700

900

12 10 8 6 4 216 1425 600 40 80 100

QRCQRC

QCQC

Q-learningQ-learning

Magnitude

of the

second

weight

vector, hh0

1

2

25 600 40 80 100 25 600 40 80 100

maxa

q(St, a)

# Steps /

episode

50

100

150

200

250

α = 2−xSteps (×103)

QCQ-learning

QC

QRC

Q-learningQRC

Steps (×103) Steps (×103)

Figure 10. Control methods on Mountain Car with neural networkfunction approximation. Each method takes one update step for ev-ery environment step and uses η = 1. Top Left: Average numberof steps to goal. Top Right: Sensitivity to stepsize showing areaunder the learning curve for each value of α. Bottom Left: Mag-nitude of the secondary weights for each algorithm. Q-learningis included as a flat line at zero, as Q-learning is effectively aspecial case of QRC where the secondary weights are always 0.Bottom Right: Mean and standard deviation of the maximumaction-value for each step of learning. QC exhibited massivegrowth in action-values throughout learning and Q-learning exhib-ited periodic spikes of instability.

QRC

QRCQC

QC

Q-learning

Q-learning QRC

Q-learning

QRC

QC

Q-learning

QC

# Steps /

episode

# Steps /

episode

200

400

600

800

400

600

800

1000

0

1

2

50

100

150

200

250

0 20 40 60 80 100

0 20 40 60 80 100 0 20 40 60 80 100

9 8 7 6 5 411 10

Magnitude

of the

second

weight

vector, hh

maxa

q(St, a)

α = 2−xSteps (×103)

Steps (×103) Steps (×103)

Figure 11. Same as Figure ?? except η = 0.5. Learning perfor-mance of QC is now competitive with Q-learning and QRC, thoughQC and Q-learning both exhibited more instability than QRC.

the TD error is high variance in the Mountain Car environ-ment, increasing the variance of the stochastic updates to thesecondary weights. We test this hypothesis by decreasingthe stepsize for the secondary weights. If the variance of theupdates is large, then a smaller stepsize can help stabilizelearning. We choose η = 1

2 and otherwise keep all otherempirical settings the same.

Figure ?? shows that QRC and QC now perform very simi-larly and only slightly outperform Q-learning. As discussedin Section 4.2, decreasing the secondary stepsize makes bothTDC and TDRC behave more similarly to TD, so this result

Q-learning

QRC

QRC

QC

Q-learning

QCmaxa

q(St, a)

QRC

QC

QC

Q-learning Q-learning

# Steps /

episode

# Steps /

episode

QRC

Steps (×103)

Steps (×103) Steps (×103)

50

100

150

200

250

0

1

2

0 5 10 15 20 25 0 5 10 15 20 25

400

600

800

400

600

800

1000

0 5 10 15 20 25α = 2−x

9 8 7 6 5 411 10

Magnitude

of the

second

weight

vector, hh

Figure 12. Same as Figure ?? except η = 0.5 and each methodtakes ten update steps for every environment step using prioritizedexperience replay.

is not surprising. Interestingly, Figure ?? shows that stillthe magnitude of the secondary weights quickly grows forQC; however, unlike the previous experiment, the secondaryweights for QRC do not quickly decay either.

Given that each of the algorithms seem to perform similarlywhen η = 1

2 , we revisit the highly off-policy experimentshown in Figure ?? when η = 1

2 . To further exaggerate theoff-policy sampling, we additionally prioritize the experi-ence replay buffer by drawing samples according to theirsquared TD error. Figure ?? shows that, while the learningcurve performance between algorithms appears to be thesame, Q-learning exhibits significant instability in its valuefunction approximation.

These preliminary experiments suggest that, like TDC, QC’sperformance is highly driven by the magnitude of its sec-ondary stepsize. When the secondary stepsize is well-tunedQC shows similar stability to QRC; while QRC remainsstable across all experimental settings. Q-learning, like TD,is sensitive to the degree of off-policy data, becoming in-creasingly unstable as more off-policy updates are made. Ineach of the experimental settings included in this section,Q-learning exhibited occasional spikes of instability; furthermotivating the desire to extend sound Gradient TD methodsfor non-linear control.

E. Additional Linear Prediction ResultsIn this section we include additional results supporting theexperiments run in the main body of the text. The primaryconclusions drawn from these results were redundant withexperiments in the text, but are included here for complete-ness.

We include results analogous to those in Section 4, exceptusing a constant stepsize on all problems. While constantstepsizes are not commonly used in practice, they are useful

Page 16: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Relative

RMSPBE

Tabular Inverted Dependent Boyan Baird

Vtrace

TDRC

TDCG

TD2

HTD Vt

race

TDRC

TDCG

TD2

TDHTD Vt

race

TDRC

TDC

GTD2

HTD TD

RC

TDC

GTD2

TDGTD2

TDRCTD

C

0.51.01.5

2.02.5

3.0

TD TD

GTD2

TDRC

TD

TDC

Vtrace

GTD2

TDRC

TD

HTD

Vtrace

GTD2

TDC

TD

TDRCTD

HTD

Vtrace

TDC

GTD2

α = 2−xα = 2−xα = 2−xα = 2−xα = 2−x

0.06

0.08

0.10

0.12

0.14

0.06

0.08

0.10

0.12

0.14

0.02

0.04

0.06

0.08

0.10

0.0

1.0

1.5

2.0

0.5

0.0

0.4

0.6

0.8

0.2

1.0

7 5 3 1 4 26 04 26 0 4 26 0 567

TDRC

HTD

TDCGTD2

TDRC

TDC

RMSPBE

Figure 13. Top: The normalized average area under the RMSPBE learning curve for each method on each problem using a constantstepsize. Each bar is normalized by TDRC’s performance so that each problem can be shown in the same range. All results are averagedover 200 independent runs with standard error bars shown at the top of each rectangle, though most are vanishingly small. Bottom:stepsize sensitivity measured using average area under the RMSPBE learning curve for each method on each problem. HTD and VTraceare not shown in Boyan’s Chain because they reduce to TD for on-policy problems. The values corresponding to the bar graphs are givenin Table 3.

RMSPBE

GTD2

TDC

TDRCTD

TDC

GTD2TDC

GTD2

TD

TDRC

HTD

HTDTDC

GTD2

-5 -3 -1 1 3 5 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5

HTD TDRC

TD

TDRCGTD2

TDRCTDC

0.12

0.06

0.08

0.10

0.20

0.30

0.35

0.25

0.400.14

0.12

0.06

0.08

0.10

0.14

0.02

0.04

0.06

0.08

0

1

2

3

4

Tabular Inverted Dependent Boyan Baird

η = 2x η = 2x η = 2x η = 2x η = 2x

Figure 14. Sensitivity to the second stepsize, for changing parameter η. All methods use a constant stepsize α. All methods are free tochoose any value of α for each specific value of η. Methods that do not have a second stepsize are shown as flat line.

for drawing clear conclusions without stepsize selectionalgorithm playing a confounding role. We show in Figure 8,that the relative performance between methods does notchange when using a constant stepsize. We do notice thatTDC performs more similarly to HTD, TD, and TDRC inthe constant stepsize case, which suggests that TDC benefitsless from using Adagrad than these other methods.

Figure 10 shows that algorithms are generally more similarin terms of stepsize sensitivity. This suggests that differ-ences in between the algorithms are less pronounced whenusing constant stepsizes, which provides more support forthe argument that empirical comparisons should simultane-ously consider modern stepsize selection algorithms.

For completeness, we include the values visualized in Fig-ure 1 as a table of values in Table 2. The standard error isreported for each entry in the table. The bold entries high-light the algorithm with the lowest RMSPBE for the givenproblem. The same is included for Figure 7 in Table ?? andfor Figure 8 in Table 3.

F. Investigating Target NetworksOne motivation for designing more stable off-policy algo-rithms is to improve learning interactions with neural net-work function approximators. A currently pervasive tech-

nique for improving stability of off-policy learning withneural networks is to use target networks. In this section, weinvestigate the impact of using target networks for each ofthe non-linear control algorithms investigated in this work.

In Figures ??, ??, and ?? we investigate the impact of syn-chronizing the target network to the value function approxi-mation after every 4, 64, and 256 updates respectively. Allthe experimental settings remain the same, other than therate of target network synchronization. The conclusionsdrawn in the main body of the paper continue to hold whenusing target networks; QC learns very slowly which is exag-gerated by increasing delay in updates to the bootstrappedtarget, QRC is stable and insensitive to choice of stepsize,and Q-learning performs well but is negatively impacted bythe introduction of target networks on these domains.

G. Parameter Settings and Other ExperimentDetails

G.1. Actor-Critic Algorithm with TDRC

We assume that the agent’s policy πθ(A|S) is parameterizedby weight vector θ. To incorporate TDRC into the one-stepactor-critic algorithm (Sutton & Barto, 2018), we simplychange the update rule for the value function approximation

Page 17: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Tabular Inverted Dependent Boyan BairdGTD2 0.079 ± 0.001 0.063 ± 0.001 0.041 ± 0.001 0.269 ± 0.003 0.357 ± 0.009TDC 0.063 ± 0.001 0.053 ± 0.001 0.034 ± 0.001 0.639 ± 0.001 0.196 ± 0.007HTD 0.048 ± 0.001 0.048 ± 0.001 0.025 ± 0.001 – 2.123 ± 0.013TD 0.046 ± 0.001 0.051 ± 0.001 0.024 ± 0.001 0.248 ± 0.003 4.101 ± 0.095

VTrace 0.060 ± 0.001 0.059 ± 0.001 0.038 ± 0.001 – 4.101 ± 0.095TDRC 0.049 ± 0.001 0.047 ± 0.001 0.026 ± 0.001 0.222 ± 0.002 0.242 ± 0.006

Table 1. Average area under the RMSPBE learning curve for each problem using the Adagrad algorithm. Bolded values highlight thelowest RMSPBE obtained for a given problem. These values correspond to the bar graphs in Figure 1.

Tabular Inverted Dependent Boyan BairdGTD2 0.094 ± 0.001 0.074 ± 0.001 0.048 ± 0.001 0.274 ± 0.006 0.356 ± 0.009TDC 0.071 ± 0.002 0.057 ± 0.001 0.033 ± 0.001 0.244 ± 0.005 0.215 ± 0.007HTD 0.060 ± 0.002 0.053 ± 0.001 0.032 ± 0.001 – 3.623 ± 0.027TD 0.058 ± 0.002 0.055 ± 0.001 0.031 ± 0.001 0.237 ± 0.006 3.993 ± 0.053

VTrace 0.069 ± 0.001 0.063 ± 0.001 0.042 ± 0.001 – 3.993 ± 0.053TDRC 0.061 ± 0.001 0.049 ± 0.001 0.031 ± 0.001 0.209 ± 0.004 0.232 ± 0.007

Table 2. Average area under the RMSPBE learning curve for each problem using the Adam stepsize selection algorithm. Bolded valueshighlight the lowest RMSPBE obtained for a given problem. These values correspond to the bar graphs in Figure 7.

Tabular Inverted Dependent Boyan BairdGTD2 0.090 ± 0.001 0.082 ± 0.001 0.044 ± 0.001 0.292 ± 0.004 0.361 ± 0.009TDC 0.075 ± 0.001 0.070 ± 0.001 0.041 ± 0.001 0.309 ± 0.004 0.205 ± 0.007HTD 0.063 ± 0.001 0.069 ± 0.002 0.035 ± 0.001 – 1184.368 ± 69.421TD 0.060 ± 0.001 0.070 ± 0.002 0.034 ± 0.001 0.226 ± 0.005 11401.550 ± 270.628

VTrace 0.072 ± 0.001 0.076 ± 0.002 0.045 ± 0.001 – 18.239 ± 0.046TDRC 0.064 ± 0.001 0.066 ± 0.001 0.036 ± 0.001 0.217 ± 0.004 0.232 ± 0.006

Table 3. Average area under the RMSPBE learning curve for each problem using the a constant stepsize. Bolded values highlight thelowest RMSPBE obtained for a given problem. These values correspond to the bar graphs in Figure 8.

Page 18: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Return

DQN

DQC

DQRC

QRC

QCQ-learning

QRC

QC

Steps(×104)

0

100

200

300

400

Return

Q-learning

DQRC

DQN

DQC

QRC

QC

Q-learning

2 60 4 8 10

QRC

QC

12 10 8 67 5

α = 2−x

100

200

300

400

Q-learning

Mountain Car Cart Pole

911

Steps(×104)

2 60 4 8 10

-200

-400

-600

-800

-1000

10 8 67 5

α = 2−x911

-900

-700

-500

-300

12

Figure 15. Non-linear control methods with target networks. Tar-get network is synchronized with the value function after every 4updates.

DQN

DQC

DQRC

QRC

QC

Q-learningQRC

QC

Steps(×104)

0

100

200

300

400

Q-learning

DQRC

DQN

DQCQRC

QC

Q-learning

2 60 4 8 10

QRC

QC

100

200

300

400

Mountain Car Cart Pole

Q-learning

12 10 8 67 5

α = 2−x91112 10 8 67 5

α = 2−x911

Steps(×104)2 60 4 8 10

-400

-600

-800

-1000

-200

-400

-600

-800

-1000

Return

Return

Figure 16. Same as Figure ??, except target network is synchro-nized after every 64 updates.

DQN

DQC

DQRC

QRC

QC

Q-learning

QRC

QC

Steps(×104)

0

100

200

300

400

Q-learning

DQRC

DQN

DQC

QRC

QC

Q-learning

2 60 4 8 10

QRC

QC

Mountain Car Cart Pole

12 10 8 67 5

α = 2−x911

100

200

300

150

250

Q-learning

12 10 8 67 5

α = 2−x911

Steps(×104)2 60 4 8 10

-400

-600

-800

-1000

-200

-400

-600

-800

Return

Return

350

Figure 17. Same as Figure ??, except target network is synchro-nized after every 256 updates.

1 2 3 4 5T T+1-1

5 4 3 2 1 T13

-3 -3 -3 -3 -3

-2 -2 -2 -2 -2 -2 02664

1000

3775

<latexit sha1_base64="6AGXF4w1i4EvlqdaORAILlHmbjc=">AAACTHichVDLSgMxFM3UV62vqks3wSK4KjNV0GXRjcsK9gGdUjKZO21oJjMkGbEM/UA3Ltz5FW5cKCKYtoOtreCBwDnn3pvcHC/mTGnbfrFyK6tr6xv5zcLW9s7uXnH/oKGiRFKo04hHsuURBZwJqGumObRiCST0ODS9wfW43rwHqVgk7vQwhk5IeoIFjBJtrG6Ruh70mEi9kGjJHkYFPIODXXde2//pmXBB+D93doslu2xPgJeJk5ESylDrFp9dP6JJCEJTTpRqO3asOymRmlEOo4KbKIgJHZAetA0VJATVSSdhjPCJcXwcRNIcofHEnZ9ISajUMPRMp9mvrxZrY/OvWjvRwWUnZSJONAg6fShIONYRHieLfSaBaj40hFDJzK6Y9okkVJv8CyYEZ/HLy6RRKTtn5crteal6lcWRR0foGJ0iB12gKrpBNVRHFD2iV/SOPqwn6836tL6mrTkrmzlEv5Bb/wZ7j6V7</latexit>

2664

0010

3775

<latexit sha1_base64="cqCJ5VxWA2CM6q16PUmD7M+7JwY=">AAACTHicfVDLSgMxFM1UrbW+Rl26CRbBVZmpgi6LblxWsA/olJLJ3GlDM5khyYhl6Ae6ceHOr3DjQhHB9IHVVjwQOPece5Ob4yecKe04z1ZuZXUtv17YKG5ube/s2nv7DRWnkkKdxjyWLZ8o4ExAXTPNoZVIIJHPoekPrsZ+8w6kYrG41cMEOhHpCRYySrSRujb1fOgxkfkR0ZLdj4p4Dgd73n+1u+jPCw9E8H1n1y45ZWcCvEzcGSmhGWpd+8kLYppGIDTlRKm26yS6kxGpGeUwKnqpgoTQAelB21BBIlCdbBLGCB8bJcBhLM0RGk/UnxMZiZQaRr7pNPv11aI3Fv/y2qkOLzoZE0mqQdDpQ2HKsY7xOFkcMAlU86EhhEpmdsW0TySh2uRfNCG4i19eJo1K2T0tV27OStXLWRwFdIiO0Aly0TmqomtUQ3VE0QN6QW/o3Xq0Xq0P63PamrNmMwfoF3L5L3tvpXs=</latexit>

2664

00

.75

.25

3775

<latexit sha1_base64="8J99paVoGwmZmC7AJC6bGatx9tI=">AAACUHicfVFLSwMxEJ6tr7q+Vj16CRbBU9mtlnoUvXisYKvQLSWbTmswm12SrFiW/kQv3vwdXjwomj58VfGDkG++mUkmX6JUcG18/9EpzM0vLC4Vl92V1bX1DW9zq6mTTDFssEQk6iqiGgWX2DDcCLxKFdI4EngZ3ZyO8pe3qDRP5IUZpNiOaV/yHmfUWKnj9cMI+1zmUUyN4ndDl3zBJ2H4X1yuVWeVSvUrDFF2P8/teCW/7I9BfpNgSkowRb3jPYTdhGUxSsME1boV+Klp51QZzgQO3TDTmFJ2Q/vYslTSGHU7HxsyJHtW6ZJeouyShozV7x05jbUexJGttPNd69ncSPwr18pM76idc5lmBiWbXNTLBDEJGblLulwhM2JgCWWK21kJu6aKMmP/wLUmBLNP/k2alXJwUK6cH5aOT6Z2FGEHdmEfAqjBMZxBHRrA4B6e4AVenQfn2XkrOJPSjx224QcK7jt8cqVy</latexit>

2664

00.5.5

3775

<latexit sha1_base64="/vA41st5smd0TLUlKOJVZuKVDTE=">AAACTnicfVHLSgMxFM3UV62vqks3wSK4KjNV0WXRjcsK9gGdoWTSO21oJjMkGbEM/UI34s7PcONCEU3bwWorHgice+4jNyd+zJnStv1s5ZaWV1bX8uuFjc2t7Z3i7l5DRYmkUKcRj2TLJwo4E1DXTHNoxRJI6HNo+oOrcb55B1KxSNzqYQxeSHqCBYwSbaROEVwfekykfki0ZPejAp7Bxq77X1w+WxBmkQui+z21UyzZZXsCvEicjJRQhlqn+OR2I5qEIDTlRKm2Y8faS4nUjHIYFdxEQUzogPSgbaggISgvndgxwkdG6eIgkuYIjSfqz46UhEoNQ99Umv36aj43Fv/KtRMdXHgpE3GiQdDpRUHCsY7w2FvcZRKo5kNDCJXM7Ippn0hCtfmBgjHBmX/yImlUys5JuXJzWqpeZnbk0QE6RMfIQeeoiq5RDdURRQ/oBb2hd+vRerU+rM9pac7KevbRL+TyX4iSpfQ=</latexit>

2664

00

.25

.75

3775

<latexit sha1_base64="AsxkF/G05HvyeLGaOLPC7GCkJEw=">AAACUHicfVFLSwMxEJ6tr7q+Vj16CRbBU9mtlnoUvXisYKvQLSWbTmswm12SrFiW/kQv3vwdXjwomj58VfGDkG++mUkmX6JUcG18/9EpzM0vLC4Vl92V1bX1DW9zq6mTTDFssEQk6iqiGgWX2DDcCLxKFdI4EngZ3ZyO8pe3qDRP5IUZpNiOaV/yHmfUWKnj9cMI+1zmUUyN4ndDl3zBJ2H4X1yuVGeVWvUrDFF2P8/teCW/7I9BfpNgSkowRb3jPYTdhGUxSsME1boV+Klp51QZzgQO3TDTmFJ2Q/vYslTSGHU7HxsyJHtW6ZJeouyShozV7x05jbUexJGttPNd69ncSPwr18pM76idc5lmBiWbXNTLBDEJGblLulwhM2JgCWWK21kJu6aKMmP/wLUmBLNP/k2alXJwUK6cH5aOT6Z2FGEHdmEfAqjBMZxBHRrA4B6e4AVenQfn2XkrOJPSjx224QcK7jt8GKVy</latexit>

2664

0001

3775

<latexit sha1_base64="iJAAx+yXnMYiCotZM/jMubGGA6U=">AAACTHiclVDLSgMxFM1UrbW+Rl26CRbBVZmpgi6LblxWsA/olJLJ3GlDM5khyYhl6Ae6ceHOr3DjQhHB9IHV1o0HAueec29yc/yEM6Ud59nKrayu5dcLG8XNre2dXXtvv6HiVFKo05jHsuUTBZwJqGumObQSCSTyOTT9wdXYb96BVCwWt3qYQCciPcFCRok2Utemng89JjI/Ilqy+1ERz+Fgz/tP7c4LD0TwfWfXLjllZwK8TNwZKaEZal37yQtimkYgNOVEqbbrJLqTEakZ5TAqeqmChNAB6UHbUEEiUJ1sEsYIHxslwGEszREaT9SfExmJlBpGvuk0+/XVojcW//LaqQ4vOhkTSapB0OlDYcqxjvE4WRwwCVTzoSGESmZ2xbRPJKHa5F80IbiLX14mjUrZPS1Xbs5K1ctZHAV0iI7QCXLROaqia1RDdUTRA3pBb+jderRerQ/rc9qas2YzB+gXcvkve1+lew==</latexit>

2664

0000

3775

<latexit sha1_base64="QU4xCQgzYKq7ocHGwzHcJcYj7aI=">AAACTHiclVDLSgMxFM1UrbW+Rl26CRbBVZmpgi6LblxWsA/olJLJ3GlDM5khyYhl6Ae6ceHOr3DjQhHB9IHV1o0HAueec29yc/yEM6Ud59nKrayu5dcLG8XNre2dXXtvv6HiVFKo05jHsuUTBZwJqGumObQSCSTyOTT9wdXYb96BVCwWt3qYQCciPcFCRok2Utemng89JjI/Ilqy+1ERz+Fgz/tXPS88EMH3nV275JSdCfAycWekhGaode0nL4hpGoHQlBOl2q6T6E5GpGaUw6jopQoSQgekB21DBYlAdbJJGCN8bJQAh7E0R2g8UX9OZCRSahj5ptPs11eL3lj8y2unOrzoZEwkqQZBpw+FKcc6xuNkccAkUM2HhhAqmdkV0z6RhGqTf9GE4C5+eZk0KmX3tFy5OStVL2dxFNAhOkInyEXnqIquUQ3VEUUP6AW9oXfr0Xq1PqzPaWvOms0coF/I5b8AecSleg==</latexit>

1 2 3 4 5 6

7

266666666664

12000000

377777777775

<latexit sha1_base64="myrdiBHVheUXkSZ1BOtEjwEKD6I=">AAACjXicrVHbSgMxEM2ut7peuuqjL8Ei+FR2q+IFkaIP+ljBXqBbSjadtqHZ7JJkxbL0b/wi3/wb0wu2tj56IHDOmRlmMhMmnCnteV+Wvba+sbmV23Z2dvf28+7BYU3FqaRQpTGPZSMkCjgTUNVMc2gkEkgUcqiHg8dxvP4GUrFYvOphAq2I9ATrMkq0sdruRxBCj4ksjIiW7H3k4Dl8HASLurSkvX/XcxWA6PwM1XYLXtGbAK8Sf0YKaIZK2/0MOjFNIxCacqJU0/cS3cqI1IxyGDlBqiAhdEB60DRUkAhUK5tsc4RPjdPB3ViaJzSeuIsVGYmUGkahyTTz9dVybGz+FWumunvdyphIUg2CTht1U451jMenwR0mgWo+NIRQycysmPaJJFSbAzpmCf7yl1dJrVT0z4ull4tC+WG2jhw6RifoDPnoCpXRM6qgKqKWY3nWjXVr5+1L+86+n6ba1qzmCP2C/fQN9vqp9w==</latexit>

266666666664

10200000

377777777775

<latexit sha1_base64="9+sgRfKAv+r3ejeBsGw+l5sAT9E=">AAACjXiclVFbS8MwFE7rbdbLpj76EhyCT6OdihdEhj7o4wR3gXWMNDubYWlaklQcZf/GX+Sb/8ZsK25ugvhB4LvkkJNzgpgzpV3307JXVtfWN3Kbztb2zm6+sLdfV1EiKdRoxCPZDIgCzgTUNNMcmrEEEgYcGsHgfpw3XkEqFolnPYyhHZK+YD1GiTZWp/DuB9BnIg1CoiV7Gzl4Bg/7/rx2F3T5j/z/eqZ8EN3vpjqFoltyJ8DLxMtIEWWodgoffjeiSQhCU06UanlurNspkZpRDiPHTxTEhA5IH1qGChKCaqeTaY7wsXG6uBdJc4TGE3e+IiWhUsMwMDdNfy9qMRubv2WtRPcu2ykTcaJB0OlDvYRjHeHxanCXSaCaDw0hVDLTK6YvRBKqzQIdMwRv8cvLpF4ueael8tNZsXKXjSOHDtEROkEeukAV9IiqqIao5ViudWVd23n73L6xb6dXbSurOUA/YD98Afbaqfc=</latexit>

266666666664

10020000

377777777775

<latexit sha1_base64="flKUnFzK5JBUDA5YBrYFqRm93TE=">AAACjXiclVHbSgMxEM2ut7peWvXRl2ARfCq7VfGCSNEHfaxgL9AtJZtOa2g2uyRZsSz9G7/IN//GtF1sbQXxQOCcMzPMZCaIOVPadT8te2V1bX0jt+lsbe/s5gt7+3UVJZJCjUY8ks2AKOBMQE0zzaEZSyBhwKERDO7H8cYrSMUi8ayHMbRD0hesxyjRxuoU3v0A+kykQUi0ZG8jB8/gYd+f1+4fuvzP/GU9Uz6I7vdQnULRLbkT4GXiZaSIMlQ7hQ+/G9EkBKEpJ0q1PDfW7ZRIzSiHkeMnCmJCB6QPLUMFCUG108k2R/jYOF3ci6R5QuOJO1+RklCpYRiYTDPfi1qMjc3fYq1E9y7bKRNxokHQaaNewrGO8Pg0uMskUM2HhhAqmZkV0xciCdXmgI5Zgrf45WVSL5e801L56axYucvWkUOH6AidIA9doAp6RFVUQ9RyLNe6sq7tvH1u39i301TbymoO0A/YD1/2uqn3</latexit>

266666666664

10002000

377777777775

<latexit sha1_base64="jrqqMZPFxUglYM+izPstx79FvHg=">AAACjXiclVHbSgMxEM2ut7peWvXRl2ARfCq7VfGCSNEHfaxgL9AtJZtOa2g2uyRZsSz9G7/IN//GtF1sbX3pgcCZM3OYyUwQc6a0635b9tr6xuZWbtvZ2d3bzxcODusqSiSFGo14JJsBUcCZgJpmmkMzlkDCgEMjGDyO8413kIpF4lUPY2iHpC9Yj1GijdQpfPoB9JlIg5BoyT5GDp7Bw74/H7srxuWV/bPIB9H9HapTKLoldwK8TLyMFFGGaqfw5XcjmoQgNOVEqZbnxrqdEqkZ5TBy/ERBTOiA9KFlqCAhqHY62eYInxqli3uRNE9oPFHnHSkJlRqGgak0872pxdxY/C/XSnTvup0yEScaBJ026iUc6wiPT4O7TALVfGgIoZKZWTF9I5JQbQ7omCV4i19eJvVyyTsvlV8uipWHbB05dIxO0Bny0BWqoGdURTVELcdyrRvr1s7bl/adfT8tta3Mc4T+wH76Afaaqfc=</latexit>

266666666664

10000200

377777777775

<latexit sha1_base64="AwTp+TH3kuknKkPbo0wXgytMGGI=">AAACjXicpVHbSgMxEM2ut7peWvXRl2ARfCq7VfGCSNEHfaxgL9AtJZtOa2g2uyRZsSz9G7/IN//GtF1sbQUfPBA458wMM5kJYs6Udt1Py15ZXVvfyG06W9s7u/nC3n5dRYmkUKMRj2QzIAo4E1DTTHNoxhJIGHBoBIP7cbzxClKxSDzrYQztkPQF6zFKtLE6hXc/gD4TaRASLdnbyMEzeNj357X7T13+M3+mfBDd76E6haJbcifAy8TLSBFlqHYKH343okkIQlNOlGp5bqzbKZGaUQ4jx08UxIQOSB9ahgoSgmqnk22O8LFxurgXSfOExhN3viIloVLDMDCZZr4XtRgbm7/FWonuXbZTJuJEg6DTRr2EYx3h8Wlwl0mgmg8NIVQyMyumL0QSqs0BHbMEb/HLy6ReLnmnpfLTWbFyl60jhw7RETpBHrpAFfSIqqiGqOVYrnVlXdt5+9y+sW+nqbaV1RygH7AfvgD2eqn3</latexit>

266666666664

10000020

377777777775

<latexit sha1_base64="nHUIpHkUqBxqkhdi+LXKxpEQUl8=">AAACjXicrVHbSgMxEM2ut7peuuqjL8Ei+FR2q+IFkaIP+ljBXqBbSjadtqHZ7JJkxbL0b/wi3/wb0wu2tj56IHDmnBlmMhMmnCnteV+Wvba+sbmV23Z2dvf28+7BYU3FqaRQpTGPZSMkCjgTUNVMc2gkEkgUcqiHg8exX38DqVgsXvUwgVZEeoJ1GSXaSG33Iwihx0QWRkRL9j5y8Bw+DoLF2PvnuLTiz6MAROdnqLZb8IreBHiV+DNSQDNU2u5n0IlpGoHQlBOlmr6X6FZGpGaUw8gJUgUJoQPSg6ahgkSgWtlkmyN8apQO7sbSPKHxRF2syEik1DAKTaaZr6+WvbH4l9dMdfe6lTGRpBoEnTbqphzrGI9PgztMAtV8aAihkplZMe0TSag2B3TMEvzlL6+SWqnonxdLLxeF8sNsHTl0jE7QGfLRFSqjZ1RBVUQtx/KsG+vWztuX9p19P021rVnNEfoF++kb9lqp9w==</latexit>

2 0 0 0 0 0 0 1

><latexit sha1_base64="TwnqhUEcgSbkxZSPTiN4QsNzzMs=">AAACRniclVBNSwMxFHxbv+tX1aOXYBE8SNmtgh5FLx4r2FboriWbvtZgNrskWbEs/XVePHvzJ3jxoIhXs20Rbb04IWSYeS95mTARXBvXfXYKM7Nz8wuLS8XlldW19dLGZkPHqWJYZ7GI1VVINQousW64EXiVKKRRKLAZ3p7lfvMOleaxvDT9BIOI9iTvckaNldqlwA+xx2UWRtQofj8gVX8/X8T95+kRH2Xn+55r38RJsV0quxV3CDJNvDEpwxi1dunJ78QsjVAaJqjWLc9NTJBRZTgTOCj6qcaEslvaw5alkkaog2wYw4DsWqVDurGyWxoyVH92ZDTSuh+FttJOeaMnvVz8y2ulpnscZFwmqUHJRg91U0FMTPJMSYcrZEb0LaFMcTsrYTdUUWZs8nkI3uSXp0mjWvEOKtWLw/LJ6TiORdiGHdgDD47gBM6hBnVg8AAv8AbvzqPz6nw4n6PSgjPu2YJfKMAXDberBg==</latexit>

Boyan’s chain Baird’s counterexample

Random walk

Figure 18. Above we provide a graphic depiction of each of thethree MDPs and the corresponding feature representations used inour experiments. We omit the three feature representations usedin the Random Walk due to space restrictions (see Sutton et al.,2009). All unlabeled transitions emit a reward of zero.

step for the TDRC update. This yields the following updateequations for Actor-Critic with TDRC:

δt = Rt+1 + γw>t xt+1 −w>t xt

wt+1 ← wt + αδtxt − γ(h>t xt)xt+1

ht+1 ← ht + ηα(δt − h>t xt

)xt − ηαβht

θt+1 ← θt + αγt+1δt∇θt lnπθ(At | St),

where the original actor-critic algorithm can be recoveredwith h0 = 0 and η = 0 and a TDC-based actor-criticalgorithm can be obtained with β = 0. In practice, theγt+1 term in the update for θ is often dropped so, as such,in our actor-critic experiment we likewise did not includethis term in our implementation. For ADAM optimizerwe used β1 = 0.9 and β2 = 0.999. We swept over α ∈2−8, 2−7, . . . , 2−2, 2−1 and had η = 1 for TDC. We usedtile coding with 5 tilings and 4× 4 tiles.

G.2. Prediction Experimental Details

For the results shown in the main body of the paper on therandom walk, Boyan’s Chain, and Baird’s Counterexamplewe swept over free meta-parameters for every method com-paring the meta-parameters which performed best accordingto the area under the RMSPBE learning curve. The step-sizes swept for all algorithms were α ∈ 2−7, 2−6, . . . , 20.For TDC and HTD, we swept values of the second step-size by sweeping over a multiplicative constant times theprimary stepsize, η ∈ 20, 21, . . . , 26 maintaining theconvergence guarantees of the two-timescale proof of con-vergence for TDC. For GTD2, we swept values of η ∈2−6, 2−5, . . . , 25, 26 as the saddlepoint formulation ofGTD2 allows for a much broader range of η while stillmaintaining convergence.

G.3. Cart Pole and Mountain Car ExperimentalDetails

To solve these task we used a fully connected neural net-work with two hidden layers where each layer had 64 nodes

Page 19: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

in Cart Pole (32 nodes in Mountain Car) with ReLU as thenon–linearity and the output layer as linear. The weightswere updated using a replay buffer of size 4,096 in CartPole (size 4000 in Mountain Car) and mini-batch size of32 using ADAM optimizer with ε = 10−8, β1 = 0.9, andβ2 = 0.999. We also used ADAM optimizer for updatingthe h vector using ε = 10−8, β1 = 0.99, and β2 = 0.999.The neural network weights were initialized using Xavierinitialization (Glorot & Bengio, 2010) and the biases wereinitialized with a normal distribution with mean 0 and stan-dard deviation 0.1. The second weight vectors were ini-tialized to 0. Actions were selected using an ε-greedy pol-icy where ε = 0.1. We tested several values of the step-size: 2−13, ..., 2−2 for Cart Pole and 2−17, ..., 2−2 forMountain Car. The final results show the performance av-eraged over 200 independent runs. In these task we setη = 1 for QC and QRC methods and set the regularizationparameter β = 1 for QRC.

G.4. MinAtar Experimental Details

We ran the MinAtar experiments for 5 million steps. Dis-count factor parameter, γ was set to 0.99. The rewards werescaled by (R× (1− γ)) so that the neural network does nothave to estimate large returns. The Q-Learning and QRCnetwork architectures were the same as that used by (Young& Tian, 2019). The network had one convolutional layerand one fully connected layer after that. The convolutionallayer used sixteen 3 × 3 convolutions with stride 1. Thefully connected layer had 128 units. Both convolutional andfully connected layers used ReLU gates. The network isinitialized the same way as (Young & Tian, 2019). We didnot use target networks for MinAtar experiments because(Young & Tian, 2019) showed that using target networkshas negligible effects on the results.

We used a circular replay buffer of size 100,000. The agentstarted learning when the replay buffer had 5,000 samplesin it. We annealed epsilon from 1.0 to 0.1 through the first100,000 steps and then kept it at 0.1 for the rest of the steps.The agent had one training step using a mini-batch of size32 per environment step. As explained by (Young & Tian,2019), frame skipping was not necessary since the frames ofthe MinAtar environment are more information rich. Otherhyperparameters were chosen the same as (Young & Tian,2019) and (Mnih et al., 2015). We used the RMSPropoptimizer with a smoothing constant of 0.95, and ε = 0.01.For QRC, we used RMSProp to learn the second weightvector h. We swept over RMSprop stepsizes in powers of 2,2−10, ..., 2−5 for breakout, and 2−12, ..., 2−8 for spaceinvaders. η was set to 1 for QC and QRC and β was 1 forQRC.

For the learning curve, we plotted the setting that resultedin the best area under the learning curve. We computed

the moving average of returns over 100 episodes (shown inFigure 5) similar to (Young & Tian, 2019). For computingthe total discounted reward, we simply averaged over all ofthe returns that the agent got during 5 million steps to geta single number for each run and each parameter setting.We then averaged this number over 30 independent runs ofthe experiment to produce one point in the bottom part ofFigure 5. For MinAtar experiments, we used python version3.7, Pytorch version 1.4, and public code made available onGithub for MinAtar1.

H. Convergence of TDRCIn this section, we prove Theorem 3.2. Our analysis closelyfollows the one timescale proof for TDC convergence (Maei,2011). We provide the full proof here for completeness.

H.1. Reformulating the TDRC Update

We combine the TDRC update equations (Eq. 8 and 9) intoa single linear system in variable %>t

def=[h>t w>t

]:

%t+1 = %t +αt(Gt+1 %t +gt+1), (12)

with Gt+1def=

[−η(xt x

>t +β I) ηρt xt(γ xt+1−xt)

>

−ρt(γ xt+1 x>t ) ρt xt(γ xt+1−xt)

>

]

and gt+1def=

[ηρtRt+1 xtρtRt+1 xt

].

For a random variable X, using the definition of importancesampling, we know that Eb[ρX] = Eπ[X]. Further, whilelearning off–policy we assume the excursion setting anduse the stationary state distribution corresponding to the be-havior policy, i.e. Eπ[xt x

>t ] =

∑S∈S db(S)x(S)x(S)>,

and consequently Eb[xt x>t ] = Eπ[xt x>t ]. Therefore, G def

=

Eb[Gk] =

[ −ηCβ −ηAA>−C −A

]and g

def= Eb[gk] =

[η bb

],

and Eq. 12 can be rewritten as

%t+1 = %t +αt(h(%t) +Mt+1

), (13)

where h(%)def= G%+g and Mt+1

def=

(Gt+1−G)%t +(gt+1−g) is the noise sequence.Also, let Ft def

= σ(%1,M1, . . . ,%t−1,Mt).

H.2. Main Proof

To prove the convergence of TDRC, we use the results fromBorkar & Meyn (2000) which require the following to betrue: (i) The function h(%) is Lipschitz and there existsh∞(%)

def= limc→∞

h(c%)c for all % ∈ R2d; (ii) The sequence

(Mt,Ft) is a Martingale difference sequence (MDS), andE[‖Mt+1‖2 | Ft

]≤ c0(1+‖% ‖2) for any initial parameter

1https://github.com/kenjyoung/MinAtar

Page 20: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Box 1: Derivation of Eq. 14.

Following the analysis given in Maei (2011), we write

det(G−λ I) = det[−ηCβ −λ I −ηA

A>−C −A−λ I

]= (−1)2d det

[ηCβ +λ I ηA

C−A> A+λ I

].

For a matrix U =

[A1 A2

A3 A4

], det(U) = det(A1) · det(A4−A3 A

−11 A2). Further, since C is positive semi–definite,

Cβ +λ I would be non–singular for any β > 0. Using these results, we get

det(G−λ I) = det(ηC+(ηβ + λ) I) · det(A+λ I−η(C−A>)(ηC+(ηβ + λ) I)−1 A). (B1)

Now ηC(ηC+(ηβ+λ) I

)−1=((ηC+(ηβ+λ) I

)−(ηβ+λ) I

)(ηC+(ηβ+λ) I

)−1= I−(ηβ+λ)

(ηC+(ηβ+

λ) I)−1

. We can then write

A+λ I−η(C−A>)(ηC+(ηβ + λ) I)−1 A

=A+λ I−ηC(ηC+(ηβ + λ) I)−1 A+ηA>(ηC+(ηβ + λ) I)−1 A

=A+λ I−(I−(ηβ + λ)

(ηC+(ηβ + λ) I

)−1)A+ηA>(ηC+(ηβ + λ) I)−1 A

=λ I+(ηβ + λ)(ηC+(ηβ + λ) I

)−1A+ηA>(ηC+(ηβ + λ) I)−1 A

=

[λ (A)

−1 (ηC+(ηβ + λ) I

)+ (ηβ + λ) I+ηA>

](ηC+(ηβ + λ) I

)−1A

= (A)−1[λ(ηC+(ηβ + λ) I

)+ A

(ηA>+(ηβ + λ) I

)](ηC+(ηβ + λ) I

)−1A .

Putting the above result in Eq. ?? along with the fact that det(A1 A2) = det(A1) · det(A2), we get

det(G−λ I) = det(λ(ηC+(ηβ + λ) I

)+ A

(ηA>+(ηβ + λ) I

)).

vector %1 and some constant c0 > 0; (iii) The stepsizesequence αt satisfies

∑t αt = ∞ and

∑t α

2t < ∞; (iv)

The origin is a globally asymptotically stable equilibriumfor the ODE % = h∞(%); and (v) The ODE % = h(%) hasa unique globally asymptotically stable equilibrium.

The function h(%) = G%+g is Lipschitz with the co-efficient ‖G ‖ and h∞(%) = G% is well defined for all% ∈ R2d. (Mt,Ft) is an MDS, since by construction it satis-fies E[Mt+1 | Ft] = 0 andMt ∈ Ft. The coverage assump-tion implies that the second moments of ρt are uniformlybounded. Then applying triangle inequality to Mt+1 =(Gt+1−G)%t +(gt+1−g) and using the boundednessof second moments of the quadruplets (xt, Rt,xt+1, ρt),we get E

[‖Mt+1‖2 | Ft

]≤ E

[‖(Gt+1−G)%t ‖2 | Ft

]+

E[‖gt+1−g ‖2 | Ft

]≤ c0(‖%t ‖2 + 1). Condition on the

stepsizes follows from our assumptions in the theorem state-ment. To verify the conditions (iv) and (v), we first showthat the real parts of all the eigenvalues of G are negative.

H.3. Proving that the Real Parts of Eigenvalues of Gare Negative (assuming C to be non–Singular)

In this section, we consider the case when the C matrix isnon–singular. TDRC converges even when C is singularunder alternate conditions, which are given in Section ??.From Box 1, we obtain

det(G−λ I) =det(λ(ηC+(ηβ + λ) I)

+ A(ηA>+(ηβ + λ) I)), (14)

for some λ ∈ C. Now because an eigenvalue λ of ma-trix G satisfies det(G−λ I) = 0, there must exist a non–zero vector z ∈ Cd such that z∗[λ(ηC+(ηβ + λ) I) +A(ηA>+(ηβ + λ) I)] z = 0, which is equivalent to

λ2 +

(ηβ + η

z∗Cz

‖ z ‖2 +z∗Az

‖ z ‖2)λ

(βz∗Az

‖ z ‖2 +z∗AA> z

‖ z ‖2

)= 0.

Page 21: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Box 2: Solutions of Eq. 15.

The solutions of a quadratic ax2 + bx+ c = 0 are given by x = − b2a ±

√b2−4ac2a . Using this, we solve for λ in Eq. 15:

2λ = −(ηβ + ηbc + λz)±√

(ηβ + ηbc + λz)2 − 4η(βλz + ba)

= −(ηβ + ηbc + (λr + λci)

)±√(

ηβ + ηbc + (λr + λci))2 − 4η

(β(λr + λci) + ba

)

= −Ω− λci±√

(Ω + λci)2 − 4η(βλr + ba)− 4ηβλci

= −Ω− λci±√(

Ω2 − λ2c − 4η(βλr + ba))

+(2Ωλc − 4ηβλc

)i

= −Ω− λci±√(

Ω2 − Ξ)

+(2Ωλc − 4ηβλc

)i,

where in the second step we put λz = λr + λci, and also we define Ω = ηβ + ηbc + λr and Ξ = λ2c + 4η(βλr + ba),which are both real numbers.

We define bc = z∗Cz‖ z ‖2 , ba = z∗AA> z

‖ z ‖2 , and λz = z∗Az‖ z ‖2 ≡

λr + λci for λr, λc ∈ R. The constants bc and ba are realand greater than zero for all non–zero vectors z. Then theabove equation can be written as

λ2 + (ηβ + ηbc + λz)λ+ η(βλz + ba) = 0. (15)

We solve for λ in Eq. 15 (see Box 2 for the full derivation) toobtain 2λ = −Ω−λci±

√(Ω2 − Ξ) + (2Ωλc − 4ηβλc)i,

where we introduced intermediate variables Ω = ηβ+ηbc+λr, and Ξ = λ2c + 4η(βλr + ba), which are both real num-bers.

Using Re(√x+ yi) = ± 1√

2

√√x2 + y2 + x we get

Re(2λ) = −Ω ± 1√2

√Υ, with the intermediate variable

Υ =√

(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2 + (Ω2 − Ξ). Nextwe obtain conditions on β and η such that the real parts ofboth the values of λ are negative for all non–zero vectorsz ∈ C.

H.3.1. CASE 1

First consider Re(2λ) = −Ω + 1√2

√Υ. Then Re(λ) < 0 is

equivalent to

Ω >1√2

√Υ. (16)

Since, the right hand side of this inequality is clearly posi-tive, we must have

Ω = ηβ + ηbc + λr > 0. (C1)

This gives us our first condition on η and β. Simplifying Eq.?? and putting back the values for the intermediate variables(see Box 3 for details), we get

Ω2 + Ξ >√

(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2. (17)

Again, since the right hand side of the above inequality ispositive, we must have

Ω2+Ξ = (ηβ+ηbc+λr)2+λ2c+4η(βλr+ba) > 0. (C2)

This is the second condition we have on η and β. Continuingto simplify the inequality in Eq. 16 (again see Box 3 fordetails), we get our third and final condition:

(ηβ+ ηbc +λr)2(βλr + ba) + βλ2c(ηbc +λr) > 0. (C3)

If λr > 0 for all z ∈ R, then each of the Conditions C1, C2,and C3 hold true and consequently TDRC converges. Thiscase corresponds to the on–policy setting where the matrixA is positive definite and TD converges.

Now we show that TDRC converges even when A is notPSD (the case where TD is not guaranteed to converge). Ifwe assume βλr + ba > 0 and ηbc + λr > 0, then each ofthe Conditions C1, C2, and C3 again hold true and TDRCwould converge. As a result we obtain the following bounds:

β < − baλr⇒ β < min

z

(−z∗AA> z

z∗Hz

), (18)

η > −λrbc⇒ η > max

z

(−z∗Hz

z∗Cz

), (19)

with Hdef= 1

2 (A+A>). These bounds can be made moreinterpretable. Using the substitution y = H

12 z we obtain

minz

(−z∗AA> z

z∗Hz

)≡ min

y

y∗(−H−12 AA>H−

12 )y

‖y ‖2

= λmin(−H−12 AA>H−

12 )

= −λmax(H−12 AA>H−

12 )

= −λmax(H−1 AA>),

where λmax represents the maximum eigenvalue of the ma-trix. Proceeding similarly for η, we can write the bounds in

Page 22: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

Box 3: Simplification of Eq. ??.

Putting the value of Υ =√

(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2 + (Ω2 − Ξ) back in Ω > 1√2

√Υ, we get

Ω >1√2

√√(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2 + (Ω2 − Ξ)

⇔ Ω2 >1

2

[√(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2 + (Ω2 − Ξ)

][squaring both sides]

⇔ Ω2 + Ξ >√

(Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2

⇔ (Ω2 + Ξ)2 > (Ω2 − Ξ)2 + (2Ωλc − 4ηβλc)2 [squaring both sides]

⇔ Ω2Ξ > (Ωλc − 2ηβλc)2

⇔ Ω2(λ2c + 4η(βλr + ba)) > Ω2λ2c + 4η2β2λ2c − 4ηβλ2cΩ [putting Ξ = λ2c + 4η(βλr + ba)]

⇔ Ω2η(βλr + ba) > η2β2λ2c − ηβλ2cΩ⇔ (ηβ + ηbc + λr)

2(βλr + ba) > ηβ2λ2c − βλ2c(ηβ + ηbc + λr) [putting Ω = ηβ + ηbc + λr]

⇔ (ηβ + ηbc + λr)2(βλr + ba) > −βλ2c(ηbc + λr)

⇔ (ηβ + ηbc + λr)2(βλr + ba) + βλ2c(ηbc + λr) > 0.

Note that all these steps have full equivalence (especially the squaring operations in second and fourth step are completelyreversible), because we explicitly enforce that Ω > 0 and Ω2 + Ξ > 0 in Conditions C1 and C2 respectively. As a result,if we satisfy conditions C1, C2, and C3, Re(2λ) = −Ω + 1√

2

√Υ < 0 would be satisfied as well.

Eq. 17 and 18 equivalently as

β < −λmax(H−1 AA>), (20)

η > −λmin(C−1 H). (21)

If these bounds are satisfied by η and β then the real partsof all the eigenvalues of G would be negative and TDRCwill converge.

H.3.2. CASE 2

Next consider Re(2λ) = −Ω − 1√2

√Υ. The second term

is always negative and we assumed Ω > 0 in Eq. C1. As aresult, Re(λ) < 0 and we are done.

Therefore, we get that the real part of the eigenvalues ofG are negative and consequently condition (iv) above issatisfied. To show that condition (v) holds true, note thatsince we assumed A+β I to be non–singular, G is alsonon–singular; this means that for the ODE % = h(%),%∗ = −G−1 g is the unique asymptotically stable equi-librium with V(%)

def= 1

2 (G%+g)>(G%+g) as its associ-ated strict Lyapunov function.

H.4. Convergence of TDRC when C is Singular

When C is singular, bc = z∗Cz‖ z ‖2 is no longer always greater

than zero for an arbitrary vector z. Consequently, if weexplicitly set bc = 0 we would get alternative bounds on ηand β for which TDRC would converge. Putting bc = 0 in

Conditions C1, C2, and C3, we get

ηβ + λr > 0,

(ηβ + λr)2 + λ2c + 4η(βλr + ba) > 0, and

(ηβ + λr)2(βλr + ba) + βλ2cλr > 0.

As before, we are concerned with the case when A is notPSD and thus λr < 0. Further, assume that βλr + ba > 0(this is the same upper bound on β as given in Eq. 17). Wesimplify the third inequality above to obtain the bound on η.As a result, we get the following bounds for β and η:

β < − baλr, η >

1

β

(√−βλ2cλrβλr + ba

− λr). (22)

The bound on η automatically satisfies the first conditionηβ + λr > 0. Therefore, if β and η satisfy these bounds,TDRC converges even for a singular C matrix.

I. Fixed Points of TDRCTheorem I.1 (Fixed Points of TDRC) If w is a TD fixedpoint, i.e., a solution to Aw = b, then it is a fixed point forthe expected TDRC update,

A>βC−1β (b−Aw) = 0.

Further, the set of fixed points for TD and TDRC are equiva-lent if Cβ is invertible and if −β does not equal to any of

Page 23: Gradient Temporal-Difference Learning with Regularized

Gradient Temporal-Difference Learning with Regularized Corrections

the eigenvalues of A. Note that Cβ is always invertible ifβ > 0, and is invertible if C is invertible even for β = 0.

Proof: To show equivalence, the first part is straight-forward: when Aw = b, then b − Aw = 0 and soA>βC

−1β (b − Aw) = 0. This means that any TD fixed

point is a TDRC fixed point. Now we simply need to showthat under the additional conditions, a TDRC fixed point isa TD fixed point.

If −β does not equal any of the eigenvalues of A, thenAβ = A + βI is a full rank matrix. Because both Aβ

and Cβ are full rank, the nullspace of A>βC−1β (b −Aw)

equals to the nullspace of b−Aw. Therefore, w satisfiesA>βC

−1β (b−Aw) = 0 iff (b−Aw) = 0.

We can prove Theorem 3.1, in an alternate fashion as well.The linear system in Eq. 12 has a solution (in expectation)which satisfies

G%+g = 0.

We show that this linear system has full rank and thus asingle solution: w = A−1 b and h = 0. If we show thatthe matrix G is non–singular, i.e. its determinant is non–zero, we are done. From Eq. 14 it is straightforward toobtain

det(G) = η2d det(A>+β I) · det(A),

which is non–zero if we assume that β does not equal thenegative of any eigenvalue of A and that A is non–singular.