πŸ“— -> 04/18/25: ECS170-L9


Games and Adversarial Search Review

🎀 Vocab

❗ Unit and Larger Context

Minimax v.s. MCTS, Which is best depends on

  • Accuracy of evaluation function for minimax or alpha-beta, versus
  • Selection policy of MCTS
    If a single move can change the course of the game, MCTS might not find it.
  • Algorithms can combine minimax and MCTS

Local Search as Optimization

  • Introduction: local search and optimization problems
  • Hill Climbing
  • Simulated Annealing
  • Beam Search
    AND-OR Search in Non-Deterministic Environments
  • Belief States and AND-OR Search

βœ’οΈ -> Scratch Notes

All problems can be formulated as search problems with varying levels of complexities. In this class we will cover the most straightforward ones

  • When path to goal doesn’t matter, you only care about the goal state
  • Iterate
    • Remember only current state
    • Move to best neighboring state, forget the past

Advantages:

  • Low memory usage
  • Often finds reasonable solutions in large/infinite state spaces
  • Useful for solving optimization problems

Optimization

Find or approximate best state according to some objective function (as the criterion for achieving the problem goal)

Usually can find an optimal solution when the search space is convex, or the objective function is convex

Optimization - Finds a global maximum or minimum
Local Search - Finds the best state (the lowest valley or the highest peak) in the state space

Hill Climbing

Given the current state , and an evaluation function , and a successor state , we want to find a reachable state that maximizes :

  • If then move to
  • Else, halt at n
    Terminates when a peak is reach
    Has no look-ahead beyond the immediate neighbors of the current state
    Chooses a random best successor if there is more than one
    Cannot backtrack, since it doesn’t remember where it’s been

How to avoid local maxima?

  • Randomly generate initial state, conduct search from there
  • Works very well with few local maxima
ProsCons
Convex problems solved fastReal problems rarely convex
If downhill moves not allowed, cannot escape local maxima
Random restart is complete, but very inefficient

Simulated Annealing

  • Combine hill climbing with a random walk in some way that yields both efficiency and completeness
  • Basic idea: allows to pick an action that results in a worse state, but gradually diminishes this probability

Adopt a control parameter T (also called β€œtemperature”)

  • T controls the probability 𝑝 in picking a worse successor
  • T starts out high to promote escape from local maxima early on
  • T gradually decreases toward 0 over time to avoid getting thrown off track late in the search (based on a β€œschedule” function)

Hill climbing only keeps track of one best state

Beam search keeps track of k states in memory

  • starts with π‘˜ randomly selected initial states
  • Find successors of all π‘˜ states
  • Take top π‘˜ successors across all successors and put them in memory
  • repeat

πŸ§ͺ -> Refresh the Info

Did you generally find the overall content understandable or compelling or relevant or not, and why, or which aspects of the reading were most novel or challenging for you and which aspects were most familiar or straightforward?)

Did a specific aspect of the reading raise questions for you or relate to other ideas and findings you’ve encountered, or are there other related issues you wish had been covered?)

Resources

  • Put useful links here

Connections

  • Link all related words