Abstract
We propose a multi-time scale quasi-Newton based smoothed functional (QN-SF) algorithm for stochastic optimization both with and without inequality constraints. The algorithm combines the smoothed functional (SF) scheme for estimating the gradient with the quasi-Newton method to solve the optimization problem. Newton algorithms typically update the Hessian at each instant and subsequently (a) project them to the space of positive definite and symmetric matrices, and (b) invert the projected Hessian. The latter operation is computationally expensive. In order to save computational effort, we propose in this paper a quasi-Newton SF (QN-SF) algorithm based on the Broyden-Fletcher-Goldfarb-Shanno (BFGS) update rule. In Bhatnagar (ACM TModel Comput S. 18(1): 27–62, 2007), a Jacobi variant of Newton SF (JN-SF) was proposed and implemented to save computational effort. We compare our QN-SF algorithm with gradient SF (G-SF) and JN-SF algorithms on two different problems – first on a simple stochastic function minimization problem and the other on a problem of optimal routing in a queueing network. We observe from the experiments that the QN-SF algorithm performs significantly better than both G-SF and JN-SF algorithms on both the problem settings. Next we extend the QN-SF algorithm to the case of constrained optimization. In this case too, the QN-SF algorithm performs much better than the JN-SF algorithm. Finally we present the proof of convergence for the QN-SF algorithm in both unconstrained and constrained settings.


Similar content being viewed by others
Notes
An alternative simpler proof of stability of \(\{Z(n)\}\) can be provided by a straightforward verification of the few stability requirements in [14] that are easily seen to hold in our setting.
References
Andradottir, S.: A scaled stochastic approximation algorithm. Manag. Sci. 42, 475–498 (1996)
Akyildiz, I.F., Su, W., Sankarasubramaniam, Y., Cayirci, E.: Wireless sensor networks: a survey. Comput. Netw. 38(4), 393–422 (2002)
Bhatnagar, S.: Adaptive multivariate three-timescale stochastic approximation algorithms for simulation based optimization. ACM Trans. Model. Comput. Simul. 15(1), 74–107 (2005)
Bhatnagar, S.: Adaptive Newton-based smoothed functional algorithms for simulation optimization. ACM Trans. Model. Comput. Simul. 18(1), 27–62 (2007)
Bhatnagar, S., Borkar, V.S.: A two time scale stochastic approximation scheme for simulation based parametric optimization. Probab. Eng. Inf. Sci. 12, 519–531 (1998)
Bhatnagar, S., Fu, M.C., Marcus, S.I., Fard, P.J.: Optimal structured feedback policies for ABR flow control using two-timescale SPSA. IEEE/ACM Trans. Netw. 9(4), 479–491 (2001)
Bhatnagar, S., Fu, M.C., Marcus, S.I., Bhatnagar, S.: Two timescale algorithms for simulation optimization of hidden Markov models. IIE Trans. 33(3), 245–258 (2001)
Bhatnagar, S., Hemachandra, N., Mishra, V.: Stochastic approximation algorithms for constrained optimization via simulation. ACM Trans. Model. Comput. Simul. 21(2), 15:1–15:22 (2011)
Bhatnagar, S., Prasad, H.L., Prashanth, L.A.: Stochastic Recursive Algorithms for Optimization: Simultaneous Perturbation Methods. Springer, New York (2013). LNCIS Series
Byrd, R.H., Hansen, S.L., Nocedal, J., Singer, Y.: A Stochastic Quasi-Newton Method for Large-Scale Optimization. CoRR arXiv:1401.7020 (2014)
Bordes, A., Bottou, L., Gallinari, P.: SGD-QN: careful quasi-Newton stochastic gradient descent. J. Mach. Learn. Res. 10, 1737–1754 (2009)
Borkar, V.S.: Stochastic Approximation: A Dynamical Systems View point. Cambridge University Press and Hindustan Book Agency, New Delhi (2008)
Borkar, V.S.: An actor-critic algorithm for constrained Markov decision processes. Syst. Control Lett. 54, 207–213 (2005)
Borkar, V.S., Meyn, S.P.: The O.D.E. method for convergence of stochastic approximation and reinforcement learning. SIAM J. Control Optim. 38(2), 447–469 (2000)
Brandiere, O.: Some pathological traps for stochastic approximation. SIAM J. Control Optim. 36, 1293–1314 (1998)
Cohen, J.E., Kelly, F.P.: A paradox of congestion in a queueing network. J. Appl. Probab. 27, 730–734 (1990)
Dennis, J.E., Morée, J.J.: Quasi-Newton methods, motivation and theory. SIAM Rev. 19(1), 46–89 (1977)
Harchol-Balter, M., Crovella, M., Murta, C.: On choosing a task assignment policy for a distributed server system. IEEE J. Parallel Distrib. Comput. 59(2), 204–228 (1999)
Hirsch, M.W.: Convergent activation dynamics in continuous time networks. Neural Netw. 2, 331–349 (1989)
Kao, C., Chen, S.: A stochastic quasi-Newton method for simulation response optimization. Eur. J. Oper. Res. 173, 30–46 (2006)
Katkovnik, V.Y., Kulchitsky, Y.: Convergence of a class of random search algorithms. Autom. Remote Control 8, 1321–1326 (1972)
Kushner, H.J., Yin, G.G.: Stochastic Approximation and Recursive Algorithms and Applications. Springer, New York (2003)
Lakshmanan, K., Bhatnagar, S.: Smoothed functional and quasi-Newton algorithms for routing in multi-stage queueing network with constraints. In: International Conference on Distributed Computing and Internet Technology (ICDCIT), vol. 6536, pp. 175–186. LNCS (2011)
Pemantle, R.: Nonconvergence to unstable points in urn models and stochastic approximations. Ann. Probab. 18, 698–712 (1990)
Schweitzer, P.J.: Perturbation theory and finite Markov chains. J. Appl. Probab. 5, 401–413 (1968)
Spall, J.C.: Adaptive stochastic approximation by the simultaneous perturbation method. IEEE Trans. Autom. Control 45, 1839–1853 (2000)
Sunehag, P., Trumpf, J., Vishwanathan, S.V.N., Schraudolph, N.N.: Variable metric stochastic approximation theory. In: Proceedings of 12th International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 560–566 (2009)
Vazquez-Abad, F.J., Kushner, H.J.: Estimation of the derivative of a stationary measure with respect to a control parameter. J. Appl. Probab. 29, 343–352 (1992)
Xiao, X., Lionel, M.N.: Internet QoS: a big picture. IEEE Netw. 13, 8–18 (1999)
Zhu, X., Spall, J.C.: A modified second-order SPSA optimization algorithm for finite samples. Int. J. Adapt. Control. 16, 397–409 (2002)
Author information
Authors and Affiliations
Corresponding author
Appendix: Proofs of Sect. 4
Appendix: Proofs of Sect. 4
Proof of Lemma 1
Note that \(Q_l(n)\) is measurable \(\mathcal {F}(n)\) for all \(n\ge 0\). Further, it is easy to see that \(E[Q_l(n+1)|\mathcal {F}(n)] = Q_l(n)\) a.s., \(\forall n \ge 0\). We can show that for any real \(a_n\) and \(b_n\),
where first we have used the Cauchy-Schwarz inequality and the second inequality follows since \(-\frac{2a_nb_n}{a_n^2+b_n^2} \le 1\). Hence we have
Here, we let \(E^a(X)\) denote \((E(X))^a\). By conditional Jensen’s inequality, we have
Hence by the Cauchy-Schwarz inequality
Since h(.) is a Lipschitz continuous function, we have \(|h(X^{'}_m) - h(X_m)|^4 \le K ||X^{'}_m - X_m ||^4\), for some constant \(K > 0\). Hence, \(E^{1/2}[(h(X^{'}_m) - h(X_m))^4] \le \sqrt{K} E^{1/2} || X^{'}_m - X_m ||^4\). As a consequence of Assumption 3, \(\sup _m E[ || X_m - X^{'}_m ||^4] < \infty \) [4]. Thus, \(E[Q^2_{l}(n)] < \infty \), for all \(n \ge 1\), i.e., \(Q_l(n)\) are square-integrable and hence also integrable random variables. Thus \((Q_l(n),\mathcal {F}(n)), n\ge 0\) is a square-integrable martingale sequence. We now show that its quadratic variation process is convergent. Thus, note that
where the second inequality follows by another application of conditional Jensen’s inequality. It can now be seen as before using an application of the Cauchy-Schwarz inequality as well as Assumptions 2 and 3 that
Now from Assumption 4, \(\sum _n E[(Q_l(n+1) - Q_l(n) )^2 | \mathcal {F}(n)] < \infty \) a.s. Thus, the quadratic variation process of \(\{Z_l(n)\}\) is almost surely convergent. Hence, by the martingale convergence theorem for square integrable martingales, \(\{Q_l(n)\}\) are a.s. convergent martingale sequences. \(\square \)
Proof of Lemma 3
Note that (9) can be rewritten as
Hence we have
Note that as a consequence of Lemma 1, \(\exists Q_l(\infty )<\infty \) a.s. such that \(Q_l(n) \rightarrow Q_l(\infty )\) a.s. as \(n\rightarrow \infty \). Now from the definition of \(Q_l(n)\), it is clear that
Note that the second term in the RHS of (35) is precisely \(Q_l(\infty )<\infty \). As a consequence of the above, it is sufficient to show the boundedness of the following recursion in place of (34):
with \({\bar{Z}}(0) = Z(0)\).
As in the proof of Lemma 1 it can be seen that
with probability 1. Now since \(b(n) \rightarrow 0\) as \(n \rightarrow \infty \), there exists a \(p_0\) such that for all \(n \ge p_0, \, 0 \le b(n) \le 1\). Hence for all \(n \ge p_0, \, {\bar{Z}}(n+1)\) is a convex combination of \({\bar{Z}}(n)\) and a quantity that is almost surely uniformly bounded. The claim followsFootnote 1.
Rights and permissions
About this article
Cite this article
Lakshmanan, K., Bhatnagar, S. Quasi-Newton smoothed functional algorithms for unconstrained and constrained simulation optimization. Comput Optim Appl 66, 533–556 (2017). https://doi.org/10.1007/s10589-016-9875-4
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10589-016-9875-4
Keywords
- Simulation
- Stochastic optimization
- Stochastic approximation algorithms
- Smoothed functional algorithm
- Quasi-Newton methods
- Constrained optimization
- Multi-stage queueing networks