2
IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS, NOVEMBER 1972 simulation. However, these errors may be small in relation to possible inaccuracies or biases in the model itself. When dealing with a questionable model, the user of the simulation should not delude himself into false confidence in the validity of his results by using a disproportionally large sample size. REFERENCES [11 G. J. Hahn and S. S. Shapiro, Statistical Models in Engineering. New York: Wiley, 1967, pp. 236-252. [2] F. S. Hillier and G. J. Lieberman, Introduction to Operations Research. San Francisco, Calif.: Holden-Day, 1967, ch. 14. [3] H. M. Wagner, Principles of Operations Research. Englewood Cliffs, N.J.: Prentice-Hall, 1969, ch. 21. [4] J. M. Hammersley and D. C. Handscomb, Monte Carlo Methods. London: Methuen, 1964. [5] T. H. Naylor, J. L. Balintfy, D. S. Burdick, and K. Chu, Computer Simulation Techniques. New York: Wiley, 1966. [6] K. D. Tocher, The Art of Simulation. New York: Van Nostrand, 1963. [7] A. E. Mace, Sample Size Determination. New York: Reinhold, 1964. [8] M. G. Natrella, Experimental Statistics, Nat. Bur. Stand. Handbook 91. Washington, D.C.: U.S. Government Printing Office, 1963. Comments on "A Stochastic Approximation Method" LENNART LJUNG Abstract-The results stated in the above paper' concerning an approved stochastic approximation method are considered. Formulas for the variances of the estimates are derived, and it is found that, in fact, the new algorithm is inferior to previously suggested ones. In the above paper' and [I ] a new stochastic approximation method is proposed. With rk, k = 1,2, , as the sequence of observations, it is suggested that x = Erk (E is the mathematical expectation) should be estimated as follows: Xn = in + Yn {fn(rn) -in-1} (1) with fn(r,n) = / ?n(l)I where n-I -n(l) = 2 rkrk+l. (2) t1 - k=1 The sequence Yn is chosen as 1/In or 1 I 1/n. The algorithms thus obtained will be referred to as algorithms I and II, respectively. Sinha and Griscik1 [1] state that these algorithms converge in the mean-square sense and that the variance of the estimate Xn is given by (.% -x) ( fl (1 - Yk)} k = 1 These algorithms will be referred to as algorithms III and IV, respectively. In Sinha and Griscik' simulations of algorithms I-IV are compared. It is also stated' that algorithms I and II converge with probability one, and a proof to this effect is given. It will be shown here that both proofs of convergence, as well as the stated variance of the estimate (3), are erroneous. However, it is true that algorithms I and II converge to jxl (and not to x) both in the mean-square sense and with probability one. Asymptotic expressions for the variances of the estimates are derived and, in contrast to the aforementioned statements' [1], it is shown that algorithms I and II are inferior to algorithm III as regards the rate of convergence. In the proof,' using Dvoretsky's theorem, the decom- position of xn+1 into the transformation of old estimates and innovation does not fulfill the elementary requirement that the innovation be independent (or virtually independent) of old estimates. The errors under discussion' [1 ] seem to stem from the fact that the estimates An(l) have not sufficiently carefully been distinguished from the mathematical expectation Erkrk+ . By iterating (1) it follows that n fcn - Ix XI = (xcl - lX 1) fl (1 Yi) i 1+1 n n + H i /i fI (I - yj) (4) i =+ 1 j =i+ I where = N/ER1(l)I - IXI. Let ri = x + cj, where sj, i = I, , is a sequence of random variables with zero-mean value and finite second and third moments, such that X and ,i+k, k > I, are independent for each i. Hence E5jX1+ I = 0. It is actually argued1 [1] that Ex1jdj+, = 0 implies that vi = Oand that thus (3) follows from (4). This is evidently false. By estimating the rate at which vi tends to zero as i -+ oo, we will, however, be able to show convergence. From (2) we obtain R(l)k k-I {x k + 1 k-+ + Rk(l) = l j+X i+ 5i+z+X k- I j=1+1 i=1 i=1 | Applying the law of the iterated logarithm [3], we see that with probability one there exists a constant M = M(co) such that 1 k < 1M. (l/2)+E j < M h- ( I > (5) (6) (3) where .*o is the initial estimate. Algorithms I and II would thus converge faster than two previously suggested algorithms (see, e.g., Tsypkin [2]). These are given by (1) with fn(rn) = rn; Yn = 1/n I n fn(rn) = - Y rk; Yn = 1 /tI. )t k=lI Manuscript received June 1, 1972. The author is with the Division of Automatic Control, Lund Institute of Technology, Lund, Sweden. ' N. K. Sinha and M. P. Griscik, IEEE Trans. Systems, Man, Cybern., vol. SMC-1, pp. 338-344, Oct. 1971. The constant M, however, depends on the realization of the process cj. Hence, if x 0, k I k-1 V/k = -~ 2 cj + 2 a:j4 + 0(k- 1+) k --> (7) k j=i 22x k j=, and Y/k = O(k- /2+E) k -+ oo, E > 0, with probability one. If x = 0 i/1k = O(k- 1/4+E), k -+ oo, with probability one. In both cases it is readily shown that the right member of (4) tends to zero as y/,i with probability one in the cases Yk = 1/k and Yk - - 1/k. Hence algorithms I and It converge with probability one to Jxl. 680

Comments on "A Stochastic Approximation Method"

  • Upload
    lennart

  • View
    215

  • Download
    2

Embed Size (px)

Citation preview

IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS, NOVEMBER 1972

simulation. However, these errors may be small in relation topossible inaccuracies or biases in the model itself. When dealingwith a questionable model, the user of the simulation should notdelude himself into false confidence in the validity of his resultsby using a disproportionally large sample size.

REFERENCES[11 G. J. Hahn and S. S. Shapiro, Statistical Models in Engineering. New

York: Wiley, 1967, pp. 236-252.[2] F. S. Hillier and G. J. Lieberman, Introduction to Operations Research.

San Francisco, Calif.: Holden-Day, 1967, ch. 14.[3] H. M. Wagner, Principles of Operations Research. Englewood Cliffs,

N.J.: Prentice-Hall, 1969, ch. 21.[4] J. M. Hammersley and D. C. Handscomb, Monte Carlo Methods.

London: Methuen, 1964.[5] T. H. Naylor, J. L. Balintfy, D. S. Burdick, and K. Chu, Computer

Simulation Techniques. New York: Wiley, 1966.[6] K. D. Tocher, The Art of Simulation. New York: Van Nostrand, 1963.[7] A. E. Mace, Sample Size Determination. New York: Reinhold, 1964.[8] M. G. Natrella, Experimental Statistics, Nat. Bur. Stand. Handbook 91.

Washington, D.C.: U.S. Government Printing Office, 1963.

Comments on "A Stochastic Approximation Method"

LENNART LJUNG

Abstract-The results stated in the above paper' concerning anapproved stochastic approximation method are considered. Formulasfor the variances of the estimates are derived, and it is found that, in fact,the new algorithm is inferior to previously suggested ones.

In the above paper' and [I ] a new stochastic approximationmethod is proposed. With rk, k = 1,2, , as the sequence ofobservations, it is suggested that x = Erk (E is the mathematicalexpectation) should be estimated as follows:

Xn = in + Yn {fn(rn) -in-1} (1)with

fn(r,n) = / ?n(l)Iwhere

n-I

-n(l)= 2 rkrk+l. (2)t1 - k=1

The sequence Yn is chosen as 1/In or 1 I1/n. The algorithms thusobtained will be referred to as algorithms I and II, respectively.Sinha and Griscik1 [1] state that these algorithms converge inthe mean-square sense and that the variance of the estimateXn is given by

(.% -x) ( fl (1 - Yk)}k= 1

These algorithms will be referred to as algorithms III and IV,respectively.

In Sinha and Griscik' simulations of algorithms I-IV arecompared. It is also stated' that algorithms I and II convergewith probability one, and a proof to this effect is given. It willbe shown here that both proofs of convergence, as well as thestated variance of the estimate (3), are erroneous. However, it istrue that algorithms I and II converge to jxl (and not to x) bothin the mean-square sense and with probability one. Asymptoticexpressions for the variances of the estimates are derived and,in contrast to the aforementioned statements' [1], it is shownthat algorithms I and II are inferior to algorithm III as regardsthe rate of convergence.

In the proof,' using Dvoretsky's theorem, the decom-position of xn+1 into the transformation of old estimatesand innovation does not fulfill the elementary requirement thatthe innovation be independent (or virtually independent) of oldestimates.The errors under discussion' [1 ] seem to stem from the fact

that the estimates An(l) have not sufficiently carefully beendistinguished from the mathematical expectation Erkrk+ .

By iterating (1) it follows thatn

fcn - IxXI = (xcl - lX 1) fl (1 Yi)i 1+1

n n

+ Hi /i fI (I - yj) (4)i=+1 j =i+ I

where

= N/ER1(l)I - IXI.Let ri = x + cj, where sj, i = I, , is a sequence of randomvariables with zero-mean value and finite second and thirdmoments, such that X and ,i+k, k > I, are independent for eachi. Hence E5jX1+ I = 0.

It is actually argued1 [1] that Ex1jdj+, = 0 implies thatvi = Oand that thus (3) follows from (4). This is evidently false.By estimating the rate at which vi tends to zero as i -+ oo, wewill, however, be able to show convergence.From (2) we obtain

R(l)kk-I{x k + 1 k-++

Rk(l) = l j+X i+ 5i+z+Xk- I j=1+1 i=1 i=1 |

Applying the law of the iterated logarithm [3], we see that withprobability one there exists a constant M = M(co) such that

1k

<1M.(l/2)+E

j < M h- ( I >(5)

(6)(3)

where .*o is the initial estimate. Algorithms I and II would thusconverge faster than two previously suggested algorithms (see,e.g., Tsypkin [2]). These are given by (1) with

fn(rn) = rn; Yn = 1/n

I nfn(rn) = - Y rk; Yn = 1 /tI.

)t k=lI

Manuscript received June 1, 1972.The author is with the Division of Automatic Control, Lund Institute

of Technology, Lund, Sweden.' N. K. Sinha and M. P. Griscik, IEEE Trans. Systems, Man, Cybern.,

vol. SMC-1, pp. 338-344, Oct. 1971.

The constant M, however, depends on the realization of theprocess cj. Hence, if x 0,

k I k-1

V/k = -~2 cj + 2 a:j4 + 0(k- 1+) k --> (7)k j=i 22x k j=,

and Y/k = O(k- /2+E) k -+ oo, E > 0, with probability one. Ifx = 0 i/1k = O(k- 1/4+E), k -+ oo, with probability one. In bothcases it is readily shown that the right member of (4) tends tozero as y/,i with probability one in the cases Yk = 1/k and Yk -- 1/k. Hence algorithms I and It converge with probability

one to Jxl.

680

CORRESPONDENCE

Using (4), it is also possible to obtain an expreasymptotic variance of the estimates. For the sake oassume that E4E2 = c2 and 1 = 1. By truncatingvariables V'k,V'j at ± IxI/2 and expanding into seric

EVIkVtiJ= max(k,j3 ( + 4x2) + 0 (max (1/k2,

With Yk = 1 l/k and x X 0,

- Ix 1)2 =-( + 4x2) + 0(1/n2),

When x = 0,

2 2G2

127rnn -* oo.

In the case Yk = l/k, we obtain

ssion for the)f definiteness;the randomas, we obtain

/j 2))

and for x # 0,

E(xn - IX 1)2

1( - 1 + 62 11+15 (2n -3)¾'2n- 2 4X2 \ 4 (nl- 1)3

+ O(/X4), x -+ 0c.k,] ° ° These variances should be compared to the one that algorithm

III yields (cr2/n) and the one that algorithm IV yields [2(I2In)].The first variance is elementary and the second is calculated in

n oo.~* (8) the same way as (10).2Thus algorithms I and II yield estimates with higher variances

than algorithm III. In fact, algorithm III provides an estimatewhose variance equals the lowest possible one, given by theCramer-Rao inequality. Any search for better algorithms inthis respect will therefore be fruitless.A plausible account for the results of the simulations given by

Sinha and Griscik' is that RX(l) has been replaced by

n k=1Hence

n n

E(fc - IxI)2 = E2 1 /In k=1 j=1

( 2 )In n= o~~ (i + n2k=1 i~ (1max (k,I.)

+ 0 (max(p i2)))

= 2u2 (I + 2) + 0 ((log n)/n2), n oo. (10)

It is also possible to obtain formulas which are not asymptoticin n. In the case when the random variable rkrk+± has normaldistribution and yn = 1, it could be shown that, for x = 0,

2a2

127r(n - 1)

1 n-IX + - Xk4k+1-

n - /k=1

Such a formula is actually given' for the estimates, but is trueonly for expected values of the estimates. The variances of theestimates obtained by such an erroneous simulation of thealgorithms are given by the second term in the right-handmembers of (8) and (10), respectively. The results shown inSinha and Griscik' are consistant with these expressions.

REFERENCES[1] N. K. Sinha and M. P. Griscik, "New algorithm for stochastic approx-

imation," IEEE Trans. Inform. Theory (Corresp.), vol. IT-17, pp.494-495, July 1971.

[2] Ya. Z. Tsypkin, Adaption and Learning in Automatic Systems. NewYork: Academic Press, 1971.

[3] K. L. Chung, A Course in Probability Theory. New York: Harcourt,Brace & World, 1968.

[4] Ya. Z. Tsypkin, "On algorithms of distribution density estimation andmoments by observations," Automat. Telemech., vol. 7, pp. 101-106,1967.

2 The expression for the variance of the estimates from algorithm IVused in Sinha and Griscik ' [1] is erroneous. This has been pointed out in [4].

681

Ecn2 =