Hill Climbing 13
-
Upload
kunwarpreet-singh -
Category
Documents
-
view
217 -
download
0
Transcript of Hill Climbing 13
![Page 1: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/1.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 1/9
![Page 2: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/2.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 2/9
It is a variant of generate and test in which the feedback
from the test procedure is used to help the generator decide
which direction to move in the search space.
In a pure generate and test procedure, the test functionresponds with only a yes or a no.
Hill climbing is often used when a good heuristic function is
available for evaluating states but when no other useful
knowledge is available..End a Hill Climbing search when there is no reasonable
alternative state to move to
![Page 3: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/3.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 3/9
` E.g. you are in an unfamiliar city without a map
` You want to go to downtown
` You simply aim for the tall buildings.
` Heuristic function : distance between the currentlocation & tall building loc.
` Desirable states will be those in which distance is
minimized
![Page 4: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/4.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 4/9
Evaluate the initial state. If it is also a goal state, then return it and
quit. Otherwise, continue with the initial state as the current state.
Loop until a solution is found or until there are no new operators left
to be applied in the current state.
Select an operator that has not yet been applied to the current state
and apply it to produce a new state.
Evaluate the new state.
If it is a goal state then return it and quit.
If it is not a goal state but it is better than the current state, then
make it the current state.
If it is not better than the current state, then continue in the loop.
![Page 5: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/5.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 5/9
` It is a variation from the simple Hill Climbing, we
consider all the moves from the current state and
select the best one possible.
` Also known as Gradient search` Both Simple Hill Climbing and Steepest Ascent may
fail to find the best solution, what kind of situations
occur are discussed on the next slide.
![Page 6: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/6.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 6/9
1. Evaluate the initial state. If it is also a goal state, thenreturn it and quit. Otherwise, continue with the initialstate as the current state.
2. Loop until a solution is found or until a complete
iteration produces no change to current state:a. Let SUCC be a state such that any possible successor of the
current state will be better than SUCC
b. For each operator that applies to the current state do:
i. Apply the operator and generate a new state
ii. Evaluate the new state. If is is a goal state, then return it and quit.
If not, compare it to SUCC. If it is better, then set SUCC to thisstate. If it is not better, leave SUCC alone.
c. If the SUCC is better than the current state, then set currentstate to SUCC,
![Page 7: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/7.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 7/9
This simple policy has three well-
known drawbacks:
1. Local Maxima: a local maximum
as opposed to global maximum.
2. Plateaus: An area of the search
space where evaluation function is
flat, thus requiring random walk.
3. Ridge: Where there are steep
slopes and the search direction is
not towards the top but towards the
side . Impossible to traverse a ridge by
single move.
(a)
(b)
(c)
Figure 5.9 Local maxima, Plateaus and
ridge situation for Hill Climbing
![Page 8: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/8.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 8/9
` In each of the previous cases (local maxima, plateaus & ridge),
the algorithm reaches a point at which no progress is being
made.
` A solution is to do a random-restart hill-climbing - where
random initial states are generated, running each until it halts
or makes no discernible progress. The best result is then
chosen.
Figure 5.10 Random-restarthill-climbing (6 initial values) for 5.9(a)
![Page 9: Hill Climbing 13](https://reader030.fdocuments.us/reader030/viewer/2022021200/577d22d81a28ab4e1e986059/html5/thumbnails/9.jpg)
8/3/2019 Hill Climbing 13
http://slidepdf.com/reader/full/hill-climbing-13 9/9
Ways to tackle these problems
Backtrack: to some earlier node & try going in a different direction.
Maintain a list of paths almost taken & go back to one of them. Deal
with the problem of local maxima.
Make a big jump: in some direction to get to a new section of search
space. Deal with the problem of plateaus.
Apply two or more rules before applying the test. This corresponds to
moving in several directions at once. Deal with ridges.