Saturday, April 6, 2019
Event Will Never Forget Essay Example for Free
 Event Will Never Forget EssayComparison of Di erent Neighbourhood Sizes in  off-key  harden Xin Yao Department of Computer Science University College, University of New South Wales Australian Defence Force  academy Canberra, ACT, Australia 2600 Abstract Neighbourhood structure and  sizing of it  be  key parameters in local   go about algorithmic programs. This is  overly true for generalised local   look for algorithms like simulated annealing. It has been shown that the  doing of simulated annealing  dismiss be improved by  dramatiseing a suitable  region  size.However, previous studies usually assumed that the  neck of the woods size was xed during  appear.     This paper presents a simulated annealing algorithm with a  kinetic  locality size which depends on the current emperature value during search. A method of dynamically deciding the  vicinity size by approximating a  sustained  probability distri only ifion is given. Four  pangtinuous probability distributions argon used in    our experiments to  perplex neighbourhood sizes dynamically, and the results are compared.  integrative optimisation.A method of generating dynamic neighbourhood sizes by approximating   hookvicttinuous probability distributions is given in this  incision. Section 4 compares the experimental results of using di erent  larntinuous probability distributions to generate dynamic neighbourhood sizes. Finally, Section 5  bustlecludes with some remarks and directions of future research. 2 General  fictitious Annealing Although SA  shadower be used in both  memorizetinuous and discrete cases, this paper only considers  integrative optimisation by SA unless  oppositewise indicated explicitly.A combinatorial optimisation  problem can be conversationally described as nding an optimum con guration X from a nite or in nite countable con guration space S . Each con guration X 2 S can be  equal by its n (gt 0) components, i. e. , X = (x1 x2  xn ), where xi 2 Xi , i = 1 2  n. An  gauzy discussion o   f combinatorial optimisation and its complexity can be found in Garey and Johnsons book 8. A general model of SA, which is applicable to both continuous and discrete problems, can be described by Figure 1, where  start generate (X Tn) is decided by the generation robability gXY (Tn ), which is the probability of generating con guration Y from con guration X at temperature Tn ,  break down  drive (X Y Tn) is decided by the acceptance probability aXY (Tn ), which is the probability of accepting con guration Y after it has been generated at temperature Tn , and  theatrical role update (Tn ) decides the rate of the temperature decrease. These  trio functions  shape the convergence of general SA 5, 6, 9, but parameters in general SA, such as the initial temperature, initial con guration, inner-loop stop criterion, and outer1 Introduction Simulated Annealing (SA) algorithms can nd  precise good near optimal solutions to a wide range of hard problems, but at the high computational cost. Va   rious methods  founder been proposed to speed up its convergence, which can roughly be divided into three categories (1) Optimising functions and parameters in SA 1 (2) Combining SA with other search algorithms 2, 3 and (3) Parallelising SA 4. This paper falls into the above rst category. Section 2 of this paper describes a general SA algorithm 5, 6 which uni es di erent variants of the classical one 7.Section 3 presents SA with a dynamic neighbourhood size and its  finish in Published in Proc. of Fourth Australian Conf. on Neural Ne iirks, ed. P. Leong and M. Jabri, pp. 216219, 1993, Melbourne, Australia. generate initial con guration X at random generate initial temperature T0 REPEAT REPEAT Y = generate(X Tn) IF accept(X Y Tn)  wherefore X = Y  UNTIL inner-loop stop criterion satis ed Tn+1 = update (Tn ) n = n + 1 UNTIL outer-loop stop criterion satis ed Figure 1 General simulated annealing. loop stop criterion, can have signi cant impact on its nite-time behaviour.That is, the co   mputation time in practice depends on the three functions as well as these parameters. Most research on SA has concentrated on the update and accept function and various algorithmic parameters, only limited attention has been paid to the generate function. However, the generate function decides an important part  the neighbourhood structure and size  of a local search algorithm irrespective of whether it is a deterministic one or a stochastic one like SA. The neighbourhood NX of a con guration X is de ned by con guration.The xed-size neighbourhood clearly does not conform with the basic search strategy behind SA. It is appealing to have a neighbourhood size which can adjust itself in the di erent search stages. Fast SA 12 can be regarded as an example of SA with a dynamic neighbourhood size, but it is only used in the continuous case. The application of dynamic neighbourhood size in combinatorial optimisation, to our best knowledge, has not been well-studied. 3 Dynamic Neighbourhood    Size in Simulated Annealing This section gives a method of dynamically deciding the neighbourhood size in SA according to the temperature parameter 5, 6.In the high temperature stages, SA algorithms have high acceptance probability for both good and ad moves, i. e. , exploration plays a major role in search, and thus a large neighbourhood size is used to  fire such exploration. In the low temperature stages, exploitation plays a major role in search, and thus a smaller neighbourhood is more suitable. In the following discussion, we say that the Hamming  distance between two con guration X = (x1  x2  xn ) and Y = (y1 y2   yn ) is if there are exactly di erent elements between them.Let f (x) be the continuous density function which is used to generate the Hamming distance between the current con guration and the next one. Denote the set of con gurations which are distant from the current con guration X as SX ( ), SX ( ) = fY 2 S gXY (Tn) gt 0g where X 62 NX , and X 2 NY i Y 2 NX . NX    = fY jY (1) ing search once de ned for a problem. Goldstein and Waterman 10 and Cheh et al. 11 carried out some experiments on comparing SA with di erent neighbourhood sizes, but the sizes are still xed once decided.A limitation of SA with a xed neighbourhood size is its inability to perform search at di erent scales in di erent stages of search. As indicated in our previous study 5, SA can be viewed as an attempt to combine exploration of a space and exploitation of a sub-space into the same algorithm, i. e. , coarse-grained search in the high temperature stages explores the con guration space and tries to locate  vivid regions, while ned-grained search in the low temperature stages exploits the promising regions and tries to nd a good near optimal gXY (Tn ) = 1=jNX j, where jNX j is the size of NX , i. . , the number of con gurations in NX , and is the same for all X in S . Moreover, jNX j is xed dur- Previous research on SA normally assumed that j Y 2 S dXY = g (2) The probabili   ty of generating con guration Y , which is dXY distant from con guration X , is dened as 1 = jS (1 )j P rob dXY ? 2 lt X dXY Z dXY + 1 2 f (x)dx = jS (1 )j 1 X dXY dXY ? 2 f (dXY ) jSX (dXY )j 2 gXY (Tn ) dXY + 1 2 (3) Suppose the maximum Hamming distance allowed for one move is dmax 1 , then the normalised generation function is f (dXY ) / jSX (dXY )j gXY (Tn ) = (4) FX (Tn ) where FX (Tn ) = X X f (d ) max XZ jSX (dXZ )j dXZ =1 Z 2S 4 Experimental Results We adopt the Traveling Salesman Problem (TSP) as a benchmark to evaluate our SA algorithms because of its clear mathematical de nition and high computational complexity. Goldstein and Waterman 10 and Cheh et al. 11 have experimented with TSPs using di erent but xed neighbourhood sizes and found that a small neighbourhood size is better than a large neighbourhood size. That is, the SA algorithm performs the best when dXY = 1. TSPs with 40 cities are used in our experiment and are generated at random.The same initial conguration, i   nner-loop stop criterion, out-loop stop criterion, and temperature decreasing rate are used in our experiments in order to evaluate the impact of the neighbourhood size on the performance of SA algorithms. Our experiments, albeit preliminary, have demonstrated that SA with a dynamic neighbourhood size outperforms SA with a xed neighbourhood size.  accede 1 gives the results of four typical runs of two kinds of SA algorithms. Table 2 gives the results of using di erent distributions to generate neighbourhood sizes. roblem instance 1 2 3 4 initial value 15080 12260 13760 15820 NorSA 2540 2140 2560 2300 CSA 3120 2520 2880 2460 Table 1 Comparison of SA with a xed neighbourhood size (CSA) and SA with a dynamic neighbourhood size (NorSA). Normal distribution is used to generate the neighbourhood size. (5) Theorem 3. 1 ( 5) Suppose the acceptance function in an SA algorithm is aXY (Tn ) = min 1 exp ? ? cY T cX n  (6) and the generation function is (4), where f (x) in (4) can be anyone of t   he following, (a) the Normal function N (0 Tn), i. e. , 1 exp ? d2 XY f (dXY ) = p 2Tn 2 Tn (b) the exponential function E (Tn ), i. . , f (dXY ) =  1 exp ? dXY Tn Tn (c) the Cauchy function C (Tn ), i. e. , 1 T f (dXY ) = 2 n 2 dXY + Tn (d) the stable function with  top executive 1 13, i. e. , 2 f (dXY ) = q exp ? 2d1 XY 2 d3 XY 1 5 Concluding Remarks Neighbourhood size is an important parameter in local search algorithms, but only a xed size was adopted in previous application of SA to combinatorial optimisation problems. This paper proposes a method of using a dynamic neighbourhood size in SA based on our analysis of SA search. Preliminary experiments have demonstrated the  emolument of a dynamic neighbourhood size in SA.The idea of a dynamic neighbourhood size could also be introduced into other local search algorithms. It is, in fact, related to a more  dense 3 Then the SA algorithm converges to  international minima if the cooling rate is Tn = ln n + n0  n = 1 2 (7) where and    n0 are positive constants. It is set to n, the number of elements in a con guration, in our experiments. 1 problem instance initial value CauSA NorSA ExpSA StableSA 1 17800 2480 2540 2640 3760 2 15500 3000 3340 3180 4420 3 16600 3300 2920 3460 4500 4 14780 3000 2980 3280 3760 References 1 P. J. M. van Laarhoven and E. H. L.Aarts, Simulated Annealing Theory and Applications, D. Reidel Publishing Co. , 1987. 2 D. H. Ackley, A Connectionist Machine for Genetic Hillclimbing, Kluwer Academic Publishers, Boston, 1987. 3 X. Yao, Optimization by  genic annealing, In M. Jabri, editor, Proc. of ACNN91, pages 9497, Sydney, 1991. 4 D. R. Greening, Parallel simulated annealing techniques, Physica D, 42293306, 1990. 5 X. Yao, Simulated annealing with extended neighbourhood,  world-wide J. of Computer Math. , 40169189, 1991. 6 X. Yao and G. -J. Li, General simulated annealing, J. of Computer Sci.  Tech. 6329 338, 1991. 7 S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi, Optimization by simulated ann   ealing, Science, 220671680, 1983. 8 M. R. Garey and D. S. Johnson, Computers and intractableness A Guide to the Theory of NP-Completeness, W. H. Freeman Co. , San Francisco, 1979. 9 S. Anily and A. Federgruen, Ergodicity in parameteric nonstationary Markov chains an application to annealing methods, Oper. Res. , 35867874, 1987. 10 L. Goldstein and M. Waterman,  neck of the woods size in the simulated annealing algorithm, Amer. J. of Math. and Management Sci. , 8409423, 1988. 11 K. M. Cheh, J.B. Goldberg, and R. G. Askin, A note on the e ect of neighborhood structure in simulated annealing algorithm, Computers and Oper. Res. , 18537547, 1991. 12 H. H. Szu and R. L. Hartley, Nonconvex optimization by  quick simulated annealing, Proc. of IEEE, 7515381540, 1987. 13 W. Feller, An Introduction to Probability Theory and Its Applications, volume 2, John Wiley  Sons, Inc. , 2nd edition, 1971. 4 Table 2 SA with a dynamic neighbourhood size which is generated by the Cauchy function (CauSA), No   rmal function (NorSA),  exponential function (ExpSA), and Stable function with index 1=2 (StableSA). esearch  case in search theory, i. e. , the issue of exploration versus exploitation or global search versus local search. Although local search based on some heuristics can be quite e cient under many circumstances, the problem of local optima is very hard to deal with. Some kind of global search has to be used if a global optimum or near optimum is required. However, the computational cost of global search is often prohibitively high for most real-world applications due to the vast search space.It is bene cial to combine global and local search together. An  dedicate question here is how to decide when global or local search should be performed. It is also di  furor to draw the line strictly between local and global search in practice. Dynamic neighbourhood size offers a way to deal with the problem by transferring from global search to local search smoothly based on a control para   meter, temperature in SA. However, more work has to be  through with(p) on deciding which kind of generation functions is most suitable for an application, i. e. what is the optimal rate of reducing the neighbourhood size. As indicated before, Fast SA 12 o ers a big improvement over classical SA 7 due to the  credence of Cauchy distribution. An interesting topic is to investigate whether the discrete version of Fast SA can o er  resembling improvement over classical SA. Our preliminary experiments seem to give a negative answer. Acknowledgement  The  write is grateful to Drs. B. Marksjo and R. Sharpe for their support of his work while he was with CSIRO Division of Building, Construction and Engineering.  
Subscribe to:
Post Comments (Atom)
 
 
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.