Richard S. Sutton, David McAllester, Satinder Singh, Yishay Mansour
AT&T Labs - Research, 180 Park Avenue, Florham Park, NJ 07932
Function approximation is essential to reinforcement learning, but the standard approach of approximating a value function and determining a policy from it has so far proven theoretically intractable. In this paper we explore an alternative approach in which the policy is explicitly represented by its own function approximator, independent of the value function, and is updated according to the gradient of expected reward with respect to the policy parameters. Williams's REINFORCE method and actor-critic methods are examples of this approach. Our main new result is to show that the gradient can be written in a form suitable for estimation from experience aided by an approximate action-value or advantage function. Using this result, we prove for the first time that a version of policy iteration with arbitrary differentiable function approximation is convergent to a locally optimal policy.
Executive Summary: Reinforcement learning (RL) enables agents, such as robots or software systems, to learn optimal behaviors by trial and error in complex environments with vast numbers of possible states. However, applying RL to real-world problems often requires function approximation—using models like neural networks to generalize from limited data—because storing exact values for every state is impossible. Traditional RL methods focus on approximating a value function that estimates long-term rewards for states or actions, then derive a policy (decision rule) by always picking the highest-value option. This approach works in many cases but falters in others: it tends to produce rigid, deterministic policies even when probabilistic ones are better, and tiny errors in value estimates can trigger drastic policy swings, preventing reliable convergence toward optimal behavior.
This paper investigates an alternative: directly approximating the policy itself with a separate function approximator, such as a neural network that outputs action probabilities based on the current state. The goal is to update this policy by following the gradient of expected long-term reward with respect to its parameters, ensuring steady improvements without relying on value-derived policies. The authors aim to establish a solid theoretical foundation for this policy gradient approach, showing how to estimate the necessary gradients from experience and proving that it converges to good solutions under function approximation.
The work relies on mathematical analysis within the standard RL framework of Markov decision processes, where an agent interacts with an environment through states, actions, and rewards. The authors derive key theorems without simulations or experiments, focusing on average reward per step or discounted rewards from a starting state. They assume differentiable policy and value approximators, bounded rewards, and that the value approximator converges to a local optimum in minimizing prediction errors. Data arises from sampling experiences under the current policy, with no specific sample sizes but general proofs applicable to large-scale MDPs.
The central result is the Policy Gradient Theorem, which states that the direction to improve policy performance equals the expected product of changes in action probabilities and the corresponding action values (long-term rewards adjusted for baselines). A major advance is showing that even an approximate action-value function—learned separately via error minimization—yields an unbiased gradient estimate if it meets a "compatibility" condition: its parameters must align with the policy's, ensuring the approximation captures relative action advantages per state rather than absolute values. This justifies using advantage functions, which highlight better-than-average actions. Finally, the authors prove that a policy iteration method—alternating value updates and policy steps along this gradient—converges to a locally optimal policy, provided step sizes decrease appropriately and gradients remain bounded.
These findings shift RL toward more stable, flexible methods that handle stochastic policies and avoid the discontinuity issues plaguing value-based approaches. By enabling direct policy optimization with approximations, the work reduces risks of divergence in high-dimensional problems, potentially speeding up learning in applications like autonomous driving or game AI, where smooth policy evolution improves safety and performance. Unlike prior results limited to special cases (e.g., tabular representations), this generalizes to arbitrary differentiable approximators, aligning with methods like REINFORCE and actor-critic algorithms that learn faster by combining policy and value updates.
Leaders should prioritize adopting policy gradient techniques in RL projects, particularly actor-critic architectures, to replace or augment value-function methods in complex domains. Start with pilots using compatible value approximations, such as linear features matching the policy, to test gains in convergence speed and policy quality. Trade-offs include higher computational needs for gradient computations versus value methods' simplicity, but benefits in reliability may outweigh this for safety-critical uses. Further analysis is needed to explore global optima and non-differentiable approximators.
Key limitations include reliance on local optima (not guaranteed best overall), assumptions of perfect value-function convergence, and untested empirical performance across diverse MDPs. Confidence is high in the theoretical proofs for the stated conditions, but readers should apply caution to unbounded or highly nonlinear real-world settings, where additional safeguards like regularization may be essential.
Large applications of reinforcement learning (RL) require the use of generalizing function approximators such as neural networks, decision-trees, or instance-based methods. The dominant approach for the last decade has been the value-function approach, in which all function approximation effort goes into estimating a value function, with the action-selection policy represented implicitly as the "greedy" policy with respect to the estimated values (e.g., as the policy that selects in each state the action with highest estimated value). The value-function approach has worked well in many applications, but has several limitations. First, it is oriented toward finding deterministic policies, whereas the optimal policy is often stochastic, selecting different actions with specific probabilities (e.g., see Singh, Jaakkola, and Jordan, 1994). Second, an arbitrarily small change in the estimated value of an action can cause it to be, or not be, selected. Such discontinuous changes have been identified as a key obstacle to establishing convergence assurances for algorithms following the value-function approach (Bertsekas and Tsitsiklis, 1996). For example, Q-learning, Sarsa, and dynamic programming methods have all been shown unable to converge to any policy for simple MDPs and simple function approximators (Gordon, 1995, 1996; Baird, 1995; Tsitsiklis and van Roy, 1996; Bertsekas and Tsitsiklis, 1996). This can occur even if the best approximation is found at each step before changing the policy, and whether the notion of "best" is in the mean-squared-error sense or the slightly different senses of residual-gradient, temporal-difference, and dynamic-programming methods.
In this paper we explore an alternative approach to function approximation in RL.
Rather than approximating a value function and using that to compute a deterministic policy, we approximate a stochastic policy directly using an independent function approximator with its own parameters. For example, the policy might be represented by a neural network whose input is a representation of the state, whose output is action selection probabilities, and whose weights are the policy parameters. Let $\theta$ denote the vector of policy parameters and $\rho$ the performance of the corresponding policy (e.g., the average reward per step). Then, in the policy gradient approach, the policy parameters are updated approximately proportional to the gradient:
$ \Delta \theta \approx \alpha \frac{\partial \rho}{\partial \theta} $
where $\alpha$ is a positive-definite step size. If the above can be achieved, then $\theta$ can usually be assured to converge to a locally optimal policy in the performance measure $\rho$. Unlike the value-function approach, here small changes in $\theta$ can cause only small changes in the policy and in the state-visitation distribution.
In this paper we prove that an unbiased estimate of the gradient (1) can be obtained from experience using an approximate value function satisfying certain properties. Williams's $(1988,1992)$ REINFORCE algorithm also finds an unbiased estimate of the gradient, but without the assistance of a learned value function. REINFORCE learns much more slowly than RL methods using value functions and has received relatively little attention. Learning a value function and using it to reduce the variance of the gradient estimate appears to be essential for rapid learning. Jaakkola, Singh and Jordan (1995) proved a result very similar to ours for the special case of function approximation corresponding to tabular POMDPs. Our result strengthens theirs and generalizes it to arbitrary differentiable function approximators. Konda and Tsitsiklis (in prep.) independently developed a very similar result to ours. See also Baxter and Bartlett (in prep.) and Marbach and Tsitsiklis (1998).
Our result also suggests a way of proving the convergence of a wide variety of algorithms based on "actor-critic" or policy-iteration architectures (e.g., Barto, Sutton, and Anderson, 1983; Sutton, 1984; Kimura and Kobayashi, 1998). In this paper we take the first step in this direction by proving for the first time that a version of policy iteration with general differentiable function approximation is convergent to a locally optimal policy. Baird and Moore (1999) obtained a weaker but superficially similar result for their VAPS family of methods. Like policy-gradient methods, VAPS includes separately parameterized policy and value functions updated by gradient methods. However, VAPS methods do not climb the gradient of performance (expected long-term reward), but of a measure combining performance and value function accuracy. As a result, VAPS does not converge to a locally optimal policy, except in the case that no weight is put upon value-function accuracy, in which case VAPS degenerates to REINFORCE. Similarly, Gordon's (1995) fitted value iteration is also convergent and value-based, but does not find a locally optimal policy.
Section Summary: In reinforcement learning, an agent learns to make decisions in an environment modeled as a Markov decision process, where it takes actions based on a policy parameterized by adjustable settings to maximize long-term rewards. The policy gradient theorem provides a formula for how to tweak these policy parameters to improve overall performance, expressed as a weighted sum involving the policy's sensitivity to changes and the expected value of actions in each state. This approach simplifies gradient estimation by avoiding direct computation of how policy changes affect state distributions, enabling practical algorithms like REINFORCE that use sampled experiences to refine the policy.
We consider the standard reinforcement learning framework (see, e.g., Sutton and Barto, 1998), in which a learning agent interacts with a Markov decision process (MDP). The state, action, and reward at each time $t \in\{0,1,2, \ldots\}$ are denoted $s_{t} \in$ $\mathcal{S}, a_{t} \in \mathcal{A}$, and $r_{t} \in \Re$ respectively. The environment's dynamics are characterized by state transition probabilities, $\mathcal{P}_{s s^{\prime}}^{a}=P r\left\{s_{t+1}=s^{\prime} \mid s_{t}=s, a_{t}=a\right\}$, and expected rewards $\mathcal{R}_{s}^{a}=E\left\{r_{t+1} \mid s_{t}=s, a_{t}=a\right\}, \forall s, s^{\prime} \in \mathcal{S}, a \in \mathcal{A}$. The agent's decision making procedure at each time is characterized by a policy, $\pi(s, a, \theta)=P r\left\{a_{t}=a \mid s_{t}=s, \theta\right\}$, $\forall s \in \mathcal{S}, a \in \mathcal{A}$, where $\theta \in \Re^{l}$, for $l<<|\mathcal{S}|$, is a parameter vector. We assume that $\pi$ is diffentiable with respect to its parameter, i.e., that $\frac{\partial \pi(s, a)}{\partial \theta}$ exists. We also usually write just $\pi(s, a)$ for $\pi(s, a, \theta)$.
With function approximation, two ways of formulating the agent's objective are useful. One is the average reward formulation, in which policies are ranked according to their long-term expected reward per step, $\rho(\pi)$ :
$ \rho(\pi)=\lim {n \rightarrow \infty} \frac{1}{n} E\left{r{1}+r_{2}+\cdots+r_{n} \mid \pi\right}=\sum_{s} d^{\pi}(s) \sum_{a} \pi(s, a) \mathcal{R}_{s}^{a} $
where $d^{\pi}(s)=\lim _{t \rightarrow \infty} \operatorname{Pr}\left\{s_{t}=s \mid s_{0}, \pi\right\}$ is the stationary distribution of states under $\pi$, which we assume exists and is independent of $s_{0}$ for all policies. In the average reward formulation, the value of a state-action pair given a policy is defined as
$ Q^{\pi}(s, a)=\sum_{t=1}^{\infty} E\left{r_{t}-\rho(\pi) \mid s_{0}=s, a_{0}=a, \pi\right}, \quad \forall s \in \mathcal{S}, a \in \mathcal{A} $
The second formulation we cover is that in which there is a designated start state $s_{0}$, and we care only about the long-term reward obtained from it. We will give our results only once, but they will apply to this formulation as well under the definitions
$ \rho(\pi)=E\left{\left.\sum_{t=1}^{\infty} \gamma^{t-1} r_{t} \right\rvert, s_{0}, \pi\right} \quad \text { and } \quad Q^{\pi}(s, a)=E\left{\left.\sum_{k=1}^{\infty} \gamma^{k-1} r_{t+k} \right\rvert, s_{t}=s, a_{t}=a, \pi\right} $
where $\gamma \in[0,1]$ is a discount rate ( $\gamma=1$ is allowed only in episodic tasks). In this formulation, we define $d^{\pi}(s)$ as a discounted weighting of states encountered starting at $s_{0}$ and then following $\pi: d^{\pi}(s)=\sum_{t=0}^{\infty} \gamma^{t} \operatorname{Pr}\left\{s_{t}=s \mid s_{0}, \pi\right\}$.
Our first result concerns the gradient of the performance metric with respect to the policy parameter:
Theorem 1 (Policy Gradient). For any MDP, in either the average-reward or start-state formulations,
$ \frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) $
Proof: See the appendix.
This way of expressing the gradient was first discussed for the average-reward formulation by Marbach and Tsitsiklis (1998), based on a related expression in terms of the state-value function due to Jaakkola, Singh, and Jordan (1995) and Cao and Chen (1997). We extend their results to the start-state formulation and provide simpler and more direct proofs. Williams's $(1988,1992)$ theory of REINFORCE algorithms can also be viewed as implying (2). In any event, the key aspect of both expressions for the gradient is that their are no terms of the form $\frac{\partial d^{\pi}(s)}{\partial \theta}$ : the effect of policy changes on the distribution of states does not appear. This is convenient for approximating the gradient by sampling. For example, if $s$ was sampled from the distribution obtained by following $\pi$, then $\sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)$ would be an unbiased estimate of $\frac{\partial \rho}{\partial \theta}$. Of course, $Q^{\pi}(s, a)$ is also not normally known and must be estimated. One approach is to use the actual returns, $R_{t}=\sum_{k=1}^{\infty} r_{t+k}-\rho(\pi)$ (or $R_{t}=\sum_{k=1}^{\infty} \gamma^{k-1} r_{t+k}$ in the start-state formulation) as an approximation for each $Q^{\pi}\left(s_{t}, a_{t}\right)$. This leads to Williams's episodic REINFORCE algorithm, $\Delta \theta_{t} \propto \frac{\partial \pi\left(s_{t}, a_{t}\right)}{\partial \theta} R_{t} \frac{1}{\pi\left(s_{t}, a_{t}\right)}$ (the $\frac{1}{\pi\left(s_{t}, a_{t}\right)}$ corrects for the oversampling of actions preferred by $\pi$ ), which is known to follow $\frac{\partial \rho}{\partial \theta}$ in expected value (Williams, 1988, 1992).
Section Summary: This section explains how to handle policy gradients in reinforcement learning when the exact Q-function, which estimates long-term rewards, is too complex to compute directly, by instead using a learned approximation called f_w. The approximation is trained by adjusting its parameters to match estimated Q-values from the current policy, and at convergence, it satisfies a condition where errors are balanced across states and actions. The main theorem proves that if this approximator is designed to align closely with the policy's own parameters—a concept called "compatible" function approximation—the gradient for improving the policy using f_w exactly equals the true gradient, ensuring the method points in the right direction for better performance.
Now consider the case in which $Q^{\pi}$ is approximated by a learned function approximator. If the approximation is sufficiently good, we might hope to use it in place of $Q^{\pi}$
in (2) and still point roughly in the direction of the gradient. For example, Jaakkola, Singh, and Jordan (1995) proved that for the special case of function approximation arising in a tabular POMDP one could assure positive inner product with the gradient, which is sufficient to ensure improvement for moving in that direction. Here we extend their result to general function approximation and prove equality with the gradient.
Let $f_{w}: \mathcal{S} \times \mathcal{A} \rightarrow \Re$ be our approximation to $Q^{\pi}$, with parameter $w$. It is natural to learn $f_{w}$ by following $\pi$ and updating $w$ by a rule such as $\Delta w_{t} \propto \frac{\partial}{\partial w}\left[\hat{Q}^{\pi}\left(s_{t}, a_{t}\right)-\right.$ $\left.f_{w}\left(s_{t}, a_{t}\right)\right]^{2} \propto\left[\hat{Q}^{\pi}\left(s_{t}, a_{t}\right)-f_{w}\left(s_{t}, a_{t}\right)\right] \frac{\partial f_{w}\left(s_{t}, a_{t}\right)}{\partial w}$, where $\hat{Q}^{\pi}\left(s_{t}, a_{t}\right)$ is some unbiased estimator of $Q^{\pi}\left(s_{t}, a_{t}\right)$, perhaps $R_{t}$. When such a process has converged to a local optimum, then
$ \sum_{s} d^{\pi}(s) \sum_{a} \pi(s, a)\left[Q^{\pi}(s, a)-f_{w}(s, a)\right] \frac{\partial f_{w}(s, a)}{\partial w}=0 $
Theorem 2 (Policy Gradient with Function Approximation). If $f_{w}$ satisfies (3) and is compatible with the policy parameterization in the sense that[^0]
$ \frac{\partial f_{w}(s, a)}{\partial w}=\frac{\partial \pi(s, a)}{\partial \theta} \frac{1}{\pi(s, a)} $
then
$ \frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} f_{w}(s, a) $
Proof: Combining (3) and (4) gives
$ \sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta}\left[Q^{\pi}(s, a)-f_{w}(s, a)\right]=0 $
which tells us that the error in $f_{w}(s, a)$ is orthogonal to the gradient of the policy parameterization. Because the expression above is zero, we can subtract it from the policy gradient theorem (2) to yield
$ \begin{aligned} \frac{\partial \rho}{\partial \theta} & =\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)-\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta}\left[Q^{\pi}(s, a)-f_{w}(s, a)\right] \ & =\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta}\left[Q^{\pi}(s, a)-Q^{\pi}(s, a)+f_{w}(s, a)\right] \ & =\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} f_{w}(s, a) \end{aligned} $
Section Summary: Theorem 2 helps derive suitable forms for approximating value functions based on how policies are parameterized, such as in a probabilistic policy using linear features for state-action pairs. For compatibility, the value function approximation must be linear in those same features but adjusted to have zero average value across actions in each state, essentially capturing the relative advantages of actions rather than absolute values. This approach justifies treating advantages as key targets in reinforcement learning, and it can be extended by adding an arbitrary state-dependent adjustment without changing the core results, though it affects the reliability of learning estimates.
Given a policy parameterization, Theorem 2 can be used to derive an appropriate form for the value-function parameterization. For example, consider a policy that is a Gibbs distribution in a linear combination of features:
$ \pi(s, a)=\frac{e^{\theta^{T} \phi_{s a}}}{\sum_{b} e^{\theta^{T} \phi_{s b}}}, \quad \forall s \in \mathcal{S}, s \in \mathcal{A} $
[^0]: Tsitsiklis (personal communication) points out that $f_{w}$ being linear in the features given on the righthand side may be the only way to satisfy this condition.
where each $\phi_{s a}$ is an $l$-dimensional feature vector characterizing state-action pair $s, a$. Meeting the compatibility condition (4) requires that
$ \frac{\partial f_{w}(s, a)}{\partial w}=\frac{\partial \pi(s, a)}{\partial \theta} \frac{1}{\pi(s, a)}=\phi_{s a}-\sum_{b} \pi(s, b) \phi_{s b} $
so that the natural parameterization of $f_{w}$ is
$ f_{w}(s, a)=w^{T}\left[\phi_{s a}-\sum_{b} \pi(s, b) \phi_{s b}\right] $
In other words, $f_{w}$ must be linear in the same features as the policy, except normalized to be mean zero for each state. Other algorithms can easily be derived for a variety of nonlinear policy parameterizations, such as multi-layer backpropagation networks.
The careful reader will have noticed that the form given above for $f_{w}$ requires that it have zero mean for each state: $\sum_{a} \pi(s, a) f_{w}(s, a)=0, \forall s \in \mathcal{S}$. In this sense it is better to think of $f_{w}$ as an approximation of the advantage function, $A^{\pi}(s, a)=Q^{\pi}(s, a)-V^{\pi}(s)$ (much as in Baird, 1993), rather than of $Q^{\pi}$. Our convergence requirement (3) is really that $f_{w}$ get the relative value of the actions correct in each state, not the absolute value, nor the variation from state to state. Our results can be viewed as a justification for the special status of advantages as the target for value function approximation in RL. In fact, our (2), (3), and (5), can all be generalized to include an arbitrary function of state added to the value function or its approximation. For example, (5) can be generalized to $\frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta}\left[f_{w}(s, a)+v(s)\right]$, where $v: \mathcal{S} \rightarrow \Re$ is an arbitrary function. (This follows immediately because $\sum_{a} \frac{\partial \pi(s, a)}{\partial \theta}=0, \forall s \in \mathcal{S}$.) The choice of $v$ does not affect any of our theorems, but can substantially affect the variance of the gradient estimators. The issues here are entirely analogous to those in the use of reinforcement baselines in earlier work (e.g., Williams, 1992; Dayan, 1991; Sutton, 1984). In practice, $v$ should presumably be set to the best available approximation of $V^{\pi}$. Our results establish that that approximation process can proceed without affecting the expected evolution of $f_{w}$ and $\pi$.
Section Summary: This section presents a theorem proving that a variant of policy iteration, which uses simplified mathematical models to approximate both policies and value functions in decision-making environments, reliably converges to a locally optimal strategy. The approach works for settings with limited rewards, where policy parameters are updated step by step in the direction that improves overall performance, guided by a compatible value estimator and diminishing but accumulating step sizes. The proof builds on earlier results to confirm that these updates follow the right path and satisfy conditions for reaching a stable optimum where further improvements flatten out.
Given Theorem 2, we can prove for the first time that a form of policy iteration with function approximation is convergent to a locally optimal policy.
Theorem 3 (Policy Iteration with Function Approximation). Let $\pi$ and $f_{w}$ be any differentiable function approximators for the policy and value function respectively that satisfy the compatibility condition (4) and for which $\max _{\theta, s, a, i, j}\left|\frac{\partial^{2} \pi(s, a)}{\partial \theta_{i} \partial \theta_{j}}\right|<B<\infty$. Let $\left\{\alpha_{k}\right\}_{k=0}^{\infty}$ be any step-size sequence such that $\lim _{k \rightarrow \infty} \alpha_{k}=0$ and $\sum_{k} \alpha_{k}=\infty$. Then, for any MDP with bounded rewards, the sequence $\left\{\rho\left(\pi_{k}\right)\right\}_{k=0}^{\infty}$, defined by any $\theta_{0}, \pi_{k}=\pi\left(\cdot, \cdot, \theta_{k}\right)$, and
$ \begin{aligned} w_{k} & =w \text { such that } \sum_{s} d^{\pi_{k}}(s) \sum_{a} \pi_{k}(s, a)\left[Q^{\pi_{k}}(s, a)-f_{w}(s, a)\right] \frac{\partial f_{w}(s, a)}{\partial w}=0 \ \theta_{k+1} & =\theta_{k}+\alpha_{k} \sum_{s} d^{\pi_{k}}(s) \sum_{a} \frac{\partial \pi_{k}(s, a)}{\partial \theta} f_{w_{k}}(s, a) \end{aligned} $
converges such that $\lim _{k \rightarrow \infty} \frac{\partial \rho\left(\pi_{k}\right)}{\partial \theta}=0$.
Proof: Our Theorem 2 assures that the $\theta_{k}$ update is in the direction of the gradient. The bounds on $\frac{\partial^{2} \pi(s, a)}{\partial \theta_{i} \partial \theta_{j}}$ and on the MDP's rewards together assure us that $\frac{\partial^{2} \rho}{\partial \theta_{i} \partial \theta_{j}}$
is also bounded. These, together with the step-size requirements, are the necessary conditions to apply Proposition 3.5 from page 96 of Bertsekas and Tsitsiklis (1996), which assures convergence to a local optimum. Q.E.D.
The authors wish to thank Martha Steenstrup and Doina Precup for comments, and Michael Kearns for insights into the notion of optimal policy under function approximation.
Section Summary: This section lists a collection of academic references on reinforcement learning, a field in artificial intelligence focused on how systems learn from trial and error to make decisions. The sources include technical reports, conference papers, journal articles, theses, and books from the 1980s and 1990s by key researchers like Andrew Barto, Richard Sutton, and John Tsitsiklis, covering topics such as algorithms for learning in uncertain environments, gradient-based methods, and function approximation techniques. These works provide foundational insights into solving complex control and optimization problems using neural networks and probabilistic models.
Baird, L. C. (1993). Advantage Updating. Wright Lab. Technical Report WL-TR-93-1146.
Baird, L. C. (1995). Residual algorithms: Reinforcement learning with function approximation. Proc. of the Twelfth Int. Conf. on Machine Learning, pp. 30-37. Morgan Kaufmann.
Baird, L. C., Moore, A. W. (1999). Gradient descent for general reinforcement learning. NIPS 11. MIT Press.
Barto, A. G., Sutton, R. S., Anderson, C. W. (1983). Neuronlike elements that can solve difficult learning control problems. IEEE Trans. on Systems, Man, and Cybernetics 13:835. Baxter, J., Bartlett, P. (in prep.) Direct gradient-based reinforcement learning: I. Gradient estimation algorithms.
Bertsekas, D. P., Tsitsiklis, J. N. (1996). Neuro-Dynamic Programming. Athena Scientific. Cao, X.-R., Chen, H.-F. (1997). Perturbation realization, potentials, and sensitivity analysis of Markov Processes, IEEE Trans. on Automatic Control 42(10):1382-1393.
Dayan, P. (1991). Reinforcement comparison. In D. S. Touretzky, J. L. Elman, T. J. Sejnowski, and G. E. Hinton (eds.), Connectionist Models: Proceedings of the 1990 Summer School, pp. 45-51. Morgan Kaufmann.
Gordon, G. J. (1995). Stable function approximation in dynamic programming. Proceedings of the Twelfth Int. Conf. on Machine Learning, pp. 261-268. Morgan Kaufmann.
Gordon, G. J. (1996). Chattering in SARSA( $\lambda$ ). CMU Learning Lab Technical Report.
Jaakkola, T., Singh, S. P., Jordan, M. I. (1995) Reinforcement learning algorithms for partially observable Markov decision problems, NIPS 7, pp. 345-352. Morgan Kaufmann.
Kimura, H., Kobayashi, S. (1998). An analysis of actor/critic algorithms using eligibility traces: Reinforcement learning with imperfect value functions. Proc. ICML-98, pp. 278-286. Konda, V. R., Tsitsiklis, J. N. (in prep.) Actor-critic algorithms.
Marbach, P., Tsitsiklis, J. N. (1998) Simulation-based optimization of Markov reward processes, technical report LIDS-P-2411, Massachusetts Institute of Technology. Singh, S. P., Jaakkola, T., Jordan, M. I. (1994). Learning without state-estimation in partially observable Markovian decision problems. Proc. ICML-94, pp. 284-292.
Sutton, R. S. (1984). Temporal Credit Assignment in Reinforcement Learning. Ph.D. thesis, University of Massachusetts, Amherst.
Sutton, R. S., Barto, A. G. (1998). Reinforcement Learning: An Introduction. MIT Press.
Tsitsiklis, J. N. Van Roy, B. (1996). Feature-based methods for large scale dynamic programming. Machine Learning 22:59-94.
Williams, R. J. (1988). Toward a theory of reinforcement-learning connectionist systems. Technical Report NU-CCS-88-3, Northeastern University, College of Computer Science.
Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning 8:229-256.
Section Summary: This appendix provides a mathematical proof of Theorem 1, which shows how changes in a policy's parameters affect its overall performance in reinforcement learning tasks. For the average-reward setup, it derives the performance gradient by expanding value functions, rewards, and state transitions, ultimately simplifying to an expectation under the policy's stationary state distribution of the policy derivative times the action-value function. The start-state version with discounting follows a similar path, unfolding the long-term value into visit probabilities to arrive at the same gradient expression, weighted by discounted state frequencies from the initial state.
We prove the theorem first for the average-reward formulation and then for the startstate formulation.
$ \begin{aligned} \frac{\partial V^{\pi}(s)}{\partial \theta} & \stackrel{\text { def }}{=} \frac{\partial}{\partial \theta} \sum_{a} \pi(s, a) Q^{\pi}(s, a) \quad \forall s \in \mathcal{S} \ & =\sum_{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \frac{\partial}{\partial \theta} Q^{\pi}(s, a)\right] \end{aligned} $
$ \begin{aligned} & =\sum_{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \frac{\partial}{\partial \theta}\left[\mathcal{R}{s}^{a}-\rho(\pi)+\sum{s^{\prime}} \mathcal{P}{s s^{\prime}}^{a} V^{\pi}\left(s^{\prime}\right)\right]\right] \ & =\sum{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a)\left[-\frac{\partial \rho}{\partial \theta}+\sum_{s^{\prime}} \mathcal{P}_{s s^{\prime}}^{a} \frac{\partial V^{\pi}\left(s^{\prime}\right)}{\partial \theta}\right]\right] \end{aligned} $
Therefore,
$ \frac{\partial \rho}{\partial \theta}=\sum_{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \sum_{s^{\prime}} \mathcal{P}_{s s^{\prime}}^{a} \frac{\partial V^{\pi}\left(s^{\prime}\right)}{\partial \theta}\right]-\frac{\partial V^{\pi}(s)}{\partial \theta} $
Summing both sides over the stationary distribution $d^{\pi}$,
$ \begin{aligned} \sum_{s} d^{\pi}(s) \frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) & +\sum_{s} d^{\pi}(s) \sum_{a} \pi(s, a) \sum_{s^{\prime}} \mathcal{P}{s s^{\prime}}^{a} \frac{\partial V^{\pi}\left(s^{\prime}\right)}{\partial \theta} \ & -\sum{s} d^{\pi}(s) \frac{\partial V^{\pi}(s)}{\partial \theta} \end{aligned} $
but since $d^{\pi}$ is stationary,
$ \begin{aligned} \sum_{s} d^{\pi}(s) \frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) & +\sum_{s^{\prime}} d^{\pi}\left(s^{\prime}\right) \frac{\partial V^{\pi}\left(s^{\prime}\right)}{\partial \theta} \ & -\sum_{s} d^{\pi}(s) \frac{\partial V^{\pi}(s)}{\partial \theta} \ \frac{\partial \rho}{\partial \theta}=\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) . & \text { Q.E.D. } \end{aligned} $
For the start-state formulation:
$ \begin{aligned} \frac{\partial V^{\pi}(s)}{\partial \theta} & \stackrel{\text { def }}{=} \frac{\partial}{\partial \theta} \sum_{a} \pi(s, a) Q^{\pi}(s, a) \quad \forall s \in \mathcal{S} \ & =\sum_{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \frac{\partial}{\partial \theta} Q^{\pi}(s, a)\right] \ & =\sum_{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \frac{\partial}{\partial \theta}\left[\mathcal{R}{s}^{a}+\sum{s^{\prime}} \gamma \mathcal{P}{s s^{\prime}}^{a} V^{\pi}\left(s^{\prime}\right)\right]\right] \ & =\sum{a}\left[\frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a)+\pi(s, a) \sum_{s^{\prime}} \gamma \mathcal{P}{s s^{\prime}}^{a} \frac{\partial}{\partial \theta} V^{\pi}\left(s^{\prime}\right)\right] \ & =\sum{x} \sum_{k=0}^{\infty} \gamma^{k} \operatorname{Pr}(s \rightarrow x, k, \pi) \sum_{a} \frac{\partial \pi(x, a)}{\partial \theta} Q^{\pi}(x, a) \end{aligned} $
after several steps of unrolling (7), where $\operatorname{Pr}(s \rightarrow x, k, \pi)$ is the probability of going from state $s$ to state $x$ in $k$ steps under policy $\pi$. It is then immediate that
$ \begin{aligned} \frac{\partial \rho}{\partial \theta} & =\frac{\partial}{\partial \theta} E\left{\sum_{t=1}^{\infty} \gamma^{t-1} r_{t} \mid s_{0}, \pi\right}=\frac{\partial}{\partial \theta} V^{\pi}\left(s_{0}\right) \ & =\sum_{s} \sum_{k=0}^{\infty} \gamma^{k} \operatorname{Pr}\left(s_{0} \rightarrow s, k, \pi\right) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) \ & =\sum_{s} d^{\pi}(s) \sum_{a} \frac{\partial \pi(s, a)}{\partial \theta} Q^{\pi}(s, a) . \quad \text { Q.E.D. } \end{aligned} $