List of Dallas Mavericks seasons: Difference between revisions
en>FrescoBot m Bot: fixing section wikilinks |
→Seasons: Record |
||
Line 1: | Line 1: | ||
{{citations missing|date=December 2009}} | |||
'''Guided Local Search''' is a [[metaheuristic]] search method. A meta-heuristic method is a method that sits on top of a [[local search (optimization)|local search algorithm]] to change its behaviour. | |||
Guided Local Search builds up penalties during a search. It uses penalties to help local search algorithms escape from local minima and plateaus. When the given local search algorithm settles in a local optimum, GLS modifies the objective function using a specific scheme (explained below). Then the local search will operate using an augmented objective function, which is designed to bring the search out of the local optimum. The key is in the way that the objective function is modified. | |||
==Overview== | |||
===Solution features=== | |||
To apply GLS, solution features must be defined for the given problem. Solution features are defined to distinguish between solutions with different characteristics, so that regions of similarity around local optima can be recognized and avoided. The choice of solution features depends on the type of problem, and also to a certain extent on the local search algorithm. For each feature <math>f_i</math> a cost function <math>c_i</math> is defined. | |||
Each feature is also associated with a penalty <math>p_i</math> (initially set to 0) to record the number of occurrences of the feature in local minima. | |||
The features and costs often come directly from the objective function. For example, in the traveling salesman problem, “whether the tour goes directly from city X to city Y” can be defined to be a feature. The distance between X and Y can be defined to be the cost. In the SAT and weighted MAX-SAT problems, the features can be “whether clause C satisfied by the current assignments”. | |||
At the implementation level, we define for each feature <math>i</math> an Indicator Function <math>I_i</math> indicating whether the feature is present in the current solution or not. <math>I_i</math> is 1 when solution <math>x</math> exhibits property <math>i</math>, 0 otherwise. | |||
===Selective penalty modifications=== | |||
GLS computes the utility of penalising each feature. When the Local Search algorithm returns a [[maxima and minima|local minimum]] x, GLS penalizes all those features (through increments to the penalty of the features) present in that solution which have maximum utility, <math>\operatorname{util}(x,i)</math>, as defined below. | |||
: <math>\operatorname{util}(x,i) = I_i(x)\frac{c_i(x)}{1+p_i}.</math> | |||
The idea is to penalise features that have high costs, although the utility of doing so decreases as the feature is penalised more and more often. | |||
===Searching through an augmented cost function=== | |||
GLS uses an augmented cost function (defined below), to allow it to guide the Local Search algorithm out of the local minimum, through penalising features present in that local minimum. The idea is to make the local minimum more costly than the surrounding search space, where these features are not present. | |||
: <math>g(x) = f(x) + \lambda a \sum_{1\leq i\leq m} I_i(x) p_i </math> | |||
The parameter λ may be used to alter the intensification of the search for solutions. A higher value for λ will result in a more diverse search, where plateaus and basins are searched more coarsely; a low value will result in a more intensive search for the solution, where the plateaus and basins in the search landscape are searched in finer detail. The coefficient <math>a</math> is used to make the penalty part of the objective function balanced relative to changes in the objective function and is problem specific. A simple heuristic for setting <math>a</math> is simply to record the average change in objective function up until the first local minimum, and then set <math>a</math> to this value divided by the number of GLS features in the problem instance. | |||
===Extensions of Guided Local Search=== | |||
[http://www.bracil.net/CSP/papers/Mills-GLS-PhD2002.pdf Mills (2002)] has described an Extended Guided Local Search (EGLS) which utilises random moves and an aspiration criterion designed specifically for penalty based schemes. The resulting algorithm improved the robustness of GLS over a range of parameter settings, particularly in the case of the [[quadratic assignment problem]]. A general version of the GLS algorithm, using a min-conflicts based hill climber (Minton et al. 1992) and based partly on GENET for constraint satisfaction and optimisation, has also been implemented in the [http://www.bracil.net/CSP/cacp/ Computer Aided Constraint Programming project]. | |||
[http://www.bracil.net/CSP/papers/Alsheddy-PhD-2011.pdf Alsheddy (2011)] extended Guided Local Search to [[multi-objective optimization]], and demonstrated its use in staff empowerment in scheduling. | |||
==Related work== | |||
GLS was built on GENET, which was developed by Chang Wang, Edward Tsang and Andrew Davenport. | |||
The [[breakout method]] is very similar to GENET. It was designed for [[constraint satisfaction]]. | |||
[[Tabu search]] is a class of search methods which can be instantiated to specific methods. GLS can be seen as a special case of [[Tabu search]]. | |||
By sitting GLS on top of genetic algorithm, Tung-leng Lau introduced the Guided Genetic Programming (GGA) algorithm. It was successfully applied to the general assignment problem (in scheduling), processors configuration problem (in electronic design) and a set of radio-link frequency assignment problems (an abstracted military application). | |||
Choi et al. cast GENET as a Lagrangian search. | |||
==Bibliography== | |||
* Alsheddy, A., Empowerment scheduling: a multi-objective optimization approach using Guided Local Search, PhD Thesis, School of Computer Science and Electronic Engineering, University of Essex, 2011 [http://www.bracil.net/CSP/papers/Alsheddy-PhD-2011.pdf] | |||
* Choi, K.M.F., Lee, J.H.M. & Stuckey, P.J., A Lagrangian Resconstruction of GENET, Artificial Intelligence, 2000, 123(1-2), 1-39 | |||
* Davenport A., Tsang E.P.K., Kangmin Zhu & C J Wang, GENET: A connectionist architecture for solving constraint satisfaction problems by iterative improvement, Proc., AAAI, 1994, p.325-330 [ftp://ftp.essex.ac.uk/pub/csp/aaai94.ps.Z] | |||
* Lau, T.L. & Tsang, E.P.K., Solving the processor configuration problem with a mutation-based genetic algorithm, International Journal on Artificial Intelligence Tools (IJAIT), World Scientific, Vol.6, No.4, December 1997, 567-585 | |||
* Lau, T.L. & Tsang, E.P.K., Guided genetic algorithm and its application to radio link frequency assignment problems, Constraints, Vol.6, No.4, 2001, 373-398 [http://www.bracil.net/CSP/papers/LauTsang-Rlfap-Constraints2001.pdf] | |||
* Lau, T.L. & Tsang, E.P.K., The guided genetic algorithm and its application to the general assignment problems, IEEE 10th International Conference on Tools with Artificial Intelligence (ICTAI'98), Taiwan, November 1998 | |||
* Mills, P. & Tsang, E.P.K., Guided local search for solving SAT and weighted MAX-SAT problems, Journal of Automated Reasoning, Special Issue on Satisfiability Problems, Kluwer, Vol.24, 2000, 205-223 [http://www.bracil.net/CSP/papers/MilTsa-Glssat-Sat2000.pdf] | |||
* Mills, P. & Tsang, E.P.K. & Ford, J., Applying an Extended Guided Local Search on the Quadratic Assignment Problem, Annals of Operations Research, Kluwer Academic Publishers, Vol.118, 2003, 121-135 [http://www.bracil.net/CSP/papers/MiTsFo-GlsQap-AnOr2002.pdf] | |||
* Minton, S., Johnston, M., Philips, A.B. & Laird, P., Minimizing conflicts: a heuristic repair method for constraint satisfaction and scheduling problems, Artificial Intelligence (Special Volume on Constraint Based Reasoning), Vol.58, Nos.1-3 1992, 161-205 | |||
* Tsang, E.P.K. & Voudouris, C., Fast local search and guided local search and their application to British Telecom's workforce scheduling problem, Operations Research Letters, Elsevier Science Publishers, Amsterdam, Vol.20, No.3, March 1997, 119-127 [http://www.bracil.net/CSP/papers/TsaVou-GlsWfs-ORL1997.pdf] | |||
* Voudouris, C, Guided local search for combinatorial optimisation problems, PhD Thesis, Department of Computer Science, University of Essex, Colchester, UK, July, 1997 [http://cswww.essex.ac.uk/staff/voudcx/doc/phd_pdf.zip] | |||
* Voudouris, C., Guided Local Search—An illustrative example in function optimisation, BT Technology Journal, Vol.16, No.3, July 1998, 46-50 [ftp://ftp.essex.ac.uk/pub/csp/Voud-GlsF6-Bttj98.pdf] | |||
* Voudouris, C. & Tsang, E.P.K., Guided Local Search and its application to the Travelling Salesman Problem, European Journal of Operational Research, Anbar Publishing, Vol.113, Issue 2, March 1999, 469-499 [http://www.bracil.net/CSP/papers/CSM-247.ps.Z] | |||
* Voudouris, C. & Tsang, E.P.K., Guided local search joins the elite in discrete optimisation, DIMACS Series in Discrete Mathematics and Theoretical Computer Science Volume 57, 2001, 29-39 [ftp://ftp.essex.ac.uk/pub/csp/VouTsa-GLSOpt-Dimacs98.ps.Z] | |||
* Voudouris, C. & Tsang, E.P.K., Guided local search, in F. Glover (ed.), Handbook of metaheuristics, Kluwer, 2003, 185-218 [http://www.bracil.net/CSP/papers/VouTsa-Gls-MetaHeuristic2003.pdf] | |||
* Voudouris, C., Tsang, E.P.K. & Alsheddy, A., Guided local search, Chapter 11, in M. Gendreau & J-Y Potvin (ed.), Handbook of Metaheuristics, Springer, 2010, 321-361 | |||
==External links== | |||
* [http://www.bracil.net/CSP/gls.html Guided Local Search Home Page] | |||
[[Category:Optimization algorithms and methods]] | |||
[[Category:Heuristics]] | |||
{{Optimization algorithms}} |
Revision as of 23:56, 2 February 2014
Template:Citations missing Guided Local Search is a metaheuristic search method. A meta-heuristic method is a method that sits on top of a local search algorithm to change its behaviour.
Guided Local Search builds up penalties during a search. It uses penalties to help local search algorithms escape from local minima and plateaus. When the given local search algorithm settles in a local optimum, GLS modifies the objective function using a specific scheme (explained below). Then the local search will operate using an augmented objective function, which is designed to bring the search out of the local optimum. The key is in the way that the objective function is modified.
Overview
Solution features
To apply GLS, solution features must be defined for the given problem. Solution features are defined to distinguish between solutions with different characteristics, so that regions of similarity around local optima can be recognized and avoided. The choice of solution features depends on the type of problem, and also to a certain extent on the local search algorithm. For each feature a cost function is defined.
Each feature is also associated with a penalty (initially set to 0) to record the number of occurrences of the feature in local minima.
The features and costs often come directly from the objective function. For example, in the traveling salesman problem, “whether the tour goes directly from city X to city Y” can be defined to be a feature. The distance between X and Y can be defined to be the cost. In the SAT and weighted MAX-SAT problems, the features can be “whether clause C satisfied by the current assignments”.
At the implementation level, we define for each feature an Indicator Function indicating whether the feature is present in the current solution or not. is 1 when solution exhibits property , 0 otherwise.
Selective penalty modifications
GLS computes the utility of penalising each feature. When the Local Search algorithm returns a local minimum x, GLS penalizes all those features (through increments to the penalty of the features) present in that solution which have maximum utility, , as defined below.
The idea is to penalise features that have high costs, although the utility of doing so decreases as the feature is penalised more and more often.
Searching through an augmented cost function
GLS uses an augmented cost function (defined below), to allow it to guide the Local Search algorithm out of the local minimum, through penalising features present in that local minimum. The idea is to make the local minimum more costly than the surrounding search space, where these features are not present.
The parameter λ may be used to alter the intensification of the search for solutions. A higher value for λ will result in a more diverse search, where plateaus and basins are searched more coarsely; a low value will result in a more intensive search for the solution, where the plateaus and basins in the search landscape are searched in finer detail. The coefficient is used to make the penalty part of the objective function balanced relative to changes in the objective function and is problem specific. A simple heuristic for setting is simply to record the average change in objective function up until the first local minimum, and then set to this value divided by the number of GLS features in the problem instance.
Extensions of Guided Local Search
Mills (2002) has described an Extended Guided Local Search (EGLS) which utilises random moves and an aspiration criterion designed specifically for penalty based schemes. The resulting algorithm improved the robustness of GLS over a range of parameter settings, particularly in the case of the quadratic assignment problem. A general version of the GLS algorithm, using a min-conflicts based hill climber (Minton et al. 1992) and based partly on GENET for constraint satisfaction and optimisation, has also been implemented in the Computer Aided Constraint Programming project.
Alsheddy (2011) extended Guided Local Search to multi-objective optimization, and demonstrated its use in staff empowerment in scheduling.
Related work
GLS was built on GENET, which was developed by Chang Wang, Edward Tsang and Andrew Davenport.
The breakout method is very similar to GENET. It was designed for constraint satisfaction.
Tabu search is a class of search methods which can be instantiated to specific methods. GLS can be seen as a special case of Tabu search.
By sitting GLS on top of genetic algorithm, Tung-leng Lau introduced the Guided Genetic Programming (GGA) algorithm. It was successfully applied to the general assignment problem (in scheduling), processors configuration problem (in electronic design) and a set of radio-link frequency assignment problems (an abstracted military application).
Choi et al. cast GENET as a Lagrangian search.
Bibliography
- Alsheddy, A., Empowerment scheduling: a multi-objective optimization approach using Guided Local Search, PhD Thesis, School of Computer Science and Electronic Engineering, University of Essex, 2011 [1]
- Choi, K.M.F., Lee, J.H.M. & Stuckey, P.J., A Lagrangian Resconstruction of GENET, Artificial Intelligence, 2000, 123(1-2), 1-39
- Davenport A., Tsang E.P.K., Kangmin Zhu & C J Wang, GENET: A connectionist architecture for solving constraint satisfaction problems by iterative improvement, Proc., AAAI, 1994, p.325-330 [2]
- Lau, T.L. & Tsang, E.P.K., Solving the processor configuration problem with a mutation-based genetic algorithm, International Journal on Artificial Intelligence Tools (IJAIT), World Scientific, Vol.6, No.4, December 1997, 567-585
- Lau, T.L. & Tsang, E.P.K., Guided genetic algorithm and its application to radio link frequency assignment problems, Constraints, Vol.6, No.4, 2001, 373-398 [3]
- Lau, T.L. & Tsang, E.P.K., The guided genetic algorithm and its application to the general assignment problems, IEEE 10th International Conference on Tools with Artificial Intelligence (ICTAI'98), Taiwan, November 1998
- Mills, P. & Tsang, E.P.K., Guided local search for solving SAT and weighted MAX-SAT problems, Journal of Automated Reasoning, Special Issue on Satisfiability Problems, Kluwer, Vol.24, 2000, 205-223 [4]
- Mills, P. & Tsang, E.P.K. & Ford, J., Applying an Extended Guided Local Search on the Quadratic Assignment Problem, Annals of Operations Research, Kluwer Academic Publishers, Vol.118, 2003, 121-135 [5]
- Minton, S., Johnston, M., Philips, A.B. & Laird, P., Minimizing conflicts: a heuristic repair method for constraint satisfaction and scheduling problems, Artificial Intelligence (Special Volume on Constraint Based Reasoning), Vol.58, Nos.1-3 1992, 161-205
- Tsang, E.P.K. & Voudouris, C., Fast local search and guided local search and their application to British Telecom's workforce scheduling problem, Operations Research Letters, Elsevier Science Publishers, Amsterdam, Vol.20, No.3, March 1997, 119-127 [6]
- Voudouris, C, Guided local search for combinatorial optimisation problems, PhD Thesis, Department of Computer Science, University of Essex, Colchester, UK, July, 1997 [7]
- Voudouris, C., Guided Local Search—An illustrative example in function optimisation, BT Technology Journal, Vol.16, No.3, July 1998, 46-50 [8]
- Voudouris, C. & Tsang, E.P.K., Guided Local Search and its application to the Travelling Salesman Problem, European Journal of Operational Research, Anbar Publishing, Vol.113, Issue 2, March 1999, 469-499 [9]
- Voudouris, C. & Tsang, E.P.K., Guided local search joins the elite in discrete optimisation, DIMACS Series in Discrete Mathematics and Theoretical Computer Science Volume 57, 2001, 29-39 [10]
- Voudouris, C. & Tsang, E.P.K., Guided local search, in F. Glover (ed.), Handbook of metaheuristics, Kluwer, 2003, 185-218 [11]
- Voudouris, C., Tsang, E.P.K. & Alsheddy, A., Guided local search, Chapter 11, in M. Gendreau & J-Y Potvin (ed.), Handbook of Metaheuristics, Springer, 2010, 321-361